| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 590 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 601 int src_index = DefineDeoptimizationLiteral(constant->handle()); | 601 int src_index = DefineDeoptimizationLiteral(constant->handle()); |
| 602 translation->StoreLiteral(src_index); | 602 translation->StoreLiteral(src_index); |
| 603 } else { | 603 } else { |
| 604 UNREACHABLE(); | 604 UNREACHABLE(); |
| 605 } | 605 } |
| 606 } | 606 } |
| 607 | 607 |
| 608 | 608 |
| 609 void LCodeGen::CallCode(Handle<Code> code, | 609 void LCodeGen::CallCode(Handle<Code> code, |
| 610 RelocInfo::Mode mode, | 610 RelocInfo::Mode mode, |
| 611 LInstruction* instr) { | 611 LInstruction* instr, |
| 612 CallCodeGeneric(code, mode, instr, RECORD_SIMPLE_SAFEPOINT); | 612 TargetAddressStorageMode storage_mode) { |
| 613 CallCodeGeneric(code, mode, instr, RECORD_SIMPLE_SAFEPOINT, storage_mode); |
| 613 } | 614 } |
| 614 | 615 |
| 615 | 616 |
| 616 void LCodeGen::CallCodeGeneric(Handle<Code> code, | 617 void LCodeGen::CallCodeGeneric(Handle<Code> code, |
| 617 RelocInfo::Mode mode, | 618 RelocInfo::Mode mode, |
| 618 LInstruction* instr, | 619 LInstruction* instr, |
| 619 SafepointMode safepoint_mode) { | 620 SafepointMode safepoint_mode, |
| 621 TargetAddressStorageMode storage_mode) { |
| 620 ASSERT(instr != NULL); | 622 ASSERT(instr != NULL); |
| 621 // Block literal pool emission to ensure nop indicating no inlined smi code | 623 // Block literal pool emission to ensure nop indicating no inlined smi code |
| 622 // is in the correct position. | 624 // is in the correct position. |
| 623 Assembler::BlockConstPoolScope block_const_pool(masm()); | 625 Assembler::BlockConstPoolScope block_const_pool(masm()); |
| 624 LPointerMap* pointers = instr->pointer_map(); | 626 LPointerMap* pointers = instr->pointer_map(); |
| 625 RecordPosition(pointers->position()); | 627 RecordPosition(pointers->position()); |
| 626 __ Call(code, mode); | 628 __ Call(code, mode, TypeFeedbackId::None(), al, storage_mode); |
| 627 RecordSafepointWithLazyDeopt(instr, safepoint_mode); | 629 RecordSafepointWithLazyDeopt(instr, safepoint_mode); |
| 628 | 630 |
| 629 // Signal that we don't inline smi code before these stubs in the | 631 // Signal that we don't inline smi code before these stubs in the |
| 630 // optimizing code generator. | 632 // optimizing code generator. |
| 631 if (code->kind() == Code::BINARY_OP_IC || | 633 if (code->kind() == Code::BINARY_OP_IC || |
| 632 code->kind() == Code::COMPARE_IC) { | 634 code->kind() == Code::COMPARE_IC) { |
| 633 __ nop(); | 635 __ nop(); |
| 634 } | 636 } |
| 635 } | 637 } |
| 636 | 638 |
| (...skipping 1827 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2464 Register map = temp; | 2466 Register map = temp; |
| 2465 __ ldr(map, FieldMemOperand(object, HeapObject::kMapOffset)); | 2467 __ ldr(map, FieldMemOperand(object, HeapObject::kMapOffset)); |
| 2466 { | 2468 { |
| 2467 // Block constant pool emission to ensure the positions of instructions are | 2469 // Block constant pool emission to ensure the positions of instructions are |
| 2468 // as expected by the patcher. See InstanceofStub::Generate(). | 2470 // as expected by the patcher. See InstanceofStub::Generate(). |
| 2469 Assembler::BlockConstPoolScope block_const_pool(masm()); | 2471 Assembler::BlockConstPoolScope block_const_pool(masm()); |
| 2470 __ bind(deferred->map_check()); // Label for calculating code patching. | 2472 __ bind(deferred->map_check()); // Label for calculating code patching. |
| 2471 // We use Factory::the_hole_value() on purpose instead of loading from the | 2473 // We use Factory::the_hole_value() on purpose instead of loading from the |
| 2472 // root array to force relocation to be able to later patch with | 2474 // root array to force relocation to be able to later patch with |
| 2473 // the cached map. | 2475 // the cached map. |
| 2476 PredictableCodeSizeScope predictable(masm_); |
| 2474 Handle<JSGlobalPropertyCell> cell = | 2477 Handle<JSGlobalPropertyCell> cell = |
| 2475 factory()->NewJSGlobalPropertyCell(factory()->the_hole_value()); | 2478 factory()->NewJSGlobalPropertyCell(factory()->the_hole_value()); |
| 2476 __ mov(ip, Operand(Handle<Object>(cell))); | 2479 __ mov(ip, Operand(Handle<Object>(cell))); |
| 2477 __ ldr(ip, FieldMemOperand(ip, JSGlobalPropertyCell::kValueOffset)); | 2480 __ ldr(ip, FieldMemOperand(ip, JSGlobalPropertyCell::kValueOffset)); |
| 2478 __ cmp(map, Operand(ip)); | 2481 __ cmp(map, Operand(ip)); |
| 2479 __ b(ne, &cache_miss); | 2482 __ b(ne, &cache_miss); |
| 2480 // We use Factory::the_hole_value() on purpose instead of loading from the | 2483 // We use Factory::the_hole_value() on purpose instead of loading from the |
| 2481 // root array to force relocation to be able to later patch | 2484 // root array to force relocation to be able to later patch |
| 2482 // with true or false. | 2485 // with true or false. |
| 2483 __ mov(result, Operand(factory()->the_hole_value())); | 2486 __ mov(result, Operand(factory()->the_hole_value())); |
| (...skipping 41 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2525 | 2528 |
| 2526 PushSafepointRegistersScope scope(this, Safepoint::kWithRegisters); | 2529 PushSafepointRegistersScope scope(this, Safepoint::kWithRegisters); |
| 2527 | 2530 |
| 2528 // Get the temp register reserved by the instruction. This needs to be r4 as | 2531 // Get the temp register reserved by the instruction. This needs to be r4 as |
| 2529 // its slot of the pushing of safepoint registers is used to communicate the | 2532 // its slot of the pushing of safepoint registers is used to communicate the |
| 2530 // offset to the location of the map check. | 2533 // offset to the location of the map check. |
| 2531 Register temp = ToRegister(instr->temp()); | 2534 Register temp = ToRegister(instr->temp()); |
| 2532 ASSERT(temp.is(r4)); | 2535 ASSERT(temp.is(r4)); |
| 2533 __ LoadHeapObject(InstanceofStub::right(), instr->function()); | 2536 __ LoadHeapObject(InstanceofStub::right(), instr->function()); |
| 2534 static const int kAdditionalDelta = 5; | 2537 static const int kAdditionalDelta = 5; |
| 2538 // Make sure that code size is predicable, since we use specific constants |
| 2539 // offsets in the code to find embedded values.. |
| 2540 PredictableCodeSizeScope predictable(masm_); |
| 2535 int delta = masm_->InstructionsGeneratedSince(map_check) + kAdditionalDelta; | 2541 int delta = masm_->InstructionsGeneratedSince(map_check) + kAdditionalDelta; |
| 2536 Label before_push_delta; | 2542 Label before_push_delta; |
| 2537 __ bind(&before_push_delta); | 2543 __ bind(&before_push_delta); |
| 2538 __ BlockConstPoolFor(kAdditionalDelta); | 2544 __ BlockConstPoolFor(kAdditionalDelta); |
| 2539 __ mov(temp, Operand(delta * kPointerSize)); | 2545 __ mov(temp, Operand(delta * kPointerSize)); |
| 2540 // The mov above can generate one or two instructions. The delta was computed | 2546 // The mov above can generate one or two instructions. The delta was computed |
| 2541 // for two instructions, so we need to pad here in case of one instruction. | 2547 // for two instructions, so we need to pad here in case of one instruction. |
| 2542 if (masm_->InstructionsGeneratedSince(&before_push_delta) != 2) { | 2548 if (masm_->InstructionsGeneratedSince(&before_push_delta) != 2) { |
| 2543 ASSERT_EQ(1, masm_->InstructionsGeneratedSince(&before_push_delta)); | 2549 ASSERT_EQ(1, masm_->InstructionsGeneratedSince(&before_push_delta)); |
| 2544 __ nop(); | 2550 __ nop(); |
| (...skipping 243 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2788 __ bind(&check_passed); | 2794 __ bind(&check_passed); |
| 2789 EmitLoadFieldOrConstantFunction( | 2795 EmitLoadFieldOrConstantFunction( |
| 2790 result, object, map, name, instr->environment()); | 2796 result, object, map, name, instr->environment()); |
| 2791 __ b(&done); | 2797 __ b(&done); |
| 2792 __ bind(&next); | 2798 __ bind(&next); |
| 2793 } | 2799 } |
| 2794 } | 2800 } |
| 2795 if (need_generic) { | 2801 if (need_generic) { |
| 2796 __ mov(r2, Operand(name)); | 2802 __ mov(r2, Operand(name)); |
| 2797 Handle<Code> ic = isolate()->builtins()->LoadIC_Initialize(); | 2803 Handle<Code> ic = isolate()->builtins()->LoadIC_Initialize(); |
| 2798 CallCode(ic, RelocInfo::CODE_TARGET, instr); | 2804 CallCode(ic, RelocInfo::CODE_TARGET, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 2799 } | 2805 } |
| 2800 __ bind(&done); | 2806 __ bind(&done); |
| 2801 } | 2807 } |
| 2802 | 2808 |
| 2803 | 2809 |
| 2804 void LCodeGen::DoLoadNamedGeneric(LLoadNamedGeneric* instr) { | 2810 void LCodeGen::DoLoadNamedGeneric(LLoadNamedGeneric* instr) { |
| 2805 ASSERT(ToRegister(instr->object()).is(r0)); | 2811 ASSERT(ToRegister(instr->object()).is(r0)); |
| 2806 ASSERT(ToRegister(instr->result()).is(r0)); | 2812 ASSERT(ToRegister(instr->result()).is(r0)); |
| 2807 | 2813 |
| 2808 // Name is always in r2. | 2814 // Name is always in r2. |
| 2809 __ mov(r2, Operand(instr->name())); | 2815 __ mov(r2, Operand(instr->name())); |
| 2810 Handle<Code> ic = isolate()->builtins()->LoadIC_Initialize(); | 2816 Handle<Code> ic = isolate()->builtins()->LoadIC_Initialize(); |
| 2811 CallCode(ic, RelocInfo::CODE_TARGET, instr); | 2817 CallCode(ic, RelocInfo::CODE_TARGET, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 2812 } | 2818 } |
| 2813 | 2819 |
| 2814 | 2820 |
| 2815 void LCodeGen::DoLoadFunctionPrototype(LLoadFunctionPrototype* instr) { | 2821 void LCodeGen::DoLoadFunctionPrototype(LLoadFunctionPrototype* instr) { |
| 2816 Register scratch = scratch0(); | 2822 Register scratch = scratch0(); |
| 2817 Register function = ToRegister(instr->function()); | 2823 Register function = ToRegister(instr->function()); |
| 2818 Register result = ToRegister(instr->result()); | 2824 Register result = ToRegister(instr->result()); |
| 2819 | 2825 |
| 2820 // Check that the function really is a function. Load map into the | 2826 // Check that the function really is a function. Load map into the |
| 2821 // result register. | 2827 // result register. |
| (...skipping 290 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3112 } | 3118 } |
| 3113 } | 3119 } |
| 3114 } | 3120 } |
| 3115 | 3121 |
| 3116 | 3122 |
| 3117 void LCodeGen::DoLoadKeyedGeneric(LLoadKeyedGeneric* instr) { | 3123 void LCodeGen::DoLoadKeyedGeneric(LLoadKeyedGeneric* instr) { |
| 3118 ASSERT(ToRegister(instr->object()).is(r1)); | 3124 ASSERT(ToRegister(instr->object()).is(r1)); |
| 3119 ASSERT(ToRegister(instr->key()).is(r0)); | 3125 ASSERT(ToRegister(instr->key()).is(r0)); |
| 3120 | 3126 |
| 3121 Handle<Code> ic = isolate()->builtins()->KeyedLoadIC_Initialize(); | 3127 Handle<Code> ic = isolate()->builtins()->KeyedLoadIC_Initialize(); |
| 3122 CallCode(ic, RelocInfo::CODE_TARGET, instr); | 3128 CallCode(ic, RelocInfo::CODE_TARGET, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 3123 } | 3129 } |
| 3124 | 3130 |
| 3125 | 3131 |
| 3126 void LCodeGen::DoArgumentsElements(LArgumentsElements* instr) { | 3132 void LCodeGen::DoArgumentsElements(LArgumentsElements* instr) { |
| 3127 Register scratch = scratch0(); | 3133 Register scratch = scratch0(); |
| 3128 Register result = ToRegister(instr->result()); | 3134 Register result = ToRegister(instr->result()); |
| 3129 | 3135 |
| 3130 if (instr->hydrogen()->from_inlined()) { | 3136 if (instr->hydrogen()->from_inlined()) { |
| 3131 __ sub(result, sp, Operand(2 * kPointerSize)); | 3137 __ sub(result, sp, Operand(2 * kPointerSize)); |
| 3132 } else { | 3138 } else { |
| (...skipping 674 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3807 } | 3813 } |
| 3808 } | 3814 } |
| 3809 | 3815 |
| 3810 | 3816 |
| 3811 void LCodeGen::DoCallKeyed(LCallKeyed* instr) { | 3817 void LCodeGen::DoCallKeyed(LCallKeyed* instr) { |
| 3812 ASSERT(ToRegister(instr->result()).is(r0)); | 3818 ASSERT(ToRegister(instr->result()).is(r0)); |
| 3813 | 3819 |
| 3814 int arity = instr->arity(); | 3820 int arity = instr->arity(); |
| 3815 Handle<Code> ic = | 3821 Handle<Code> ic = |
| 3816 isolate()->stub_cache()->ComputeKeyedCallInitialize(arity); | 3822 isolate()->stub_cache()->ComputeKeyedCallInitialize(arity); |
| 3817 CallCode(ic, RelocInfo::CODE_TARGET, instr); | 3823 CallCode(ic, RelocInfo::CODE_TARGET, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 3818 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); | 3824 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); |
| 3819 } | 3825 } |
| 3820 | 3826 |
| 3821 | 3827 |
| 3822 void LCodeGen::DoCallNamed(LCallNamed* instr) { | 3828 void LCodeGen::DoCallNamed(LCallNamed* instr) { |
| 3823 ASSERT(ToRegister(instr->result()).is(r0)); | 3829 ASSERT(ToRegister(instr->result()).is(r0)); |
| 3824 | 3830 |
| 3825 int arity = instr->arity(); | 3831 int arity = instr->arity(); |
| 3826 RelocInfo::Mode mode = RelocInfo::CODE_TARGET; | 3832 RelocInfo::Mode mode = RelocInfo::CODE_TARGET; |
| 3827 Handle<Code> ic = | 3833 Handle<Code> ic = |
| 3828 isolate()->stub_cache()->ComputeCallInitialize(arity, mode); | 3834 isolate()->stub_cache()->ComputeCallInitialize(arity, mode); |
| 3829 __ mov(r2, Operand(instr->name())); | 3835 __ mov(r2, Operand(instr->name())); |
| 3830 CallCode(ic, mode, instr); | 3836 CallCode(ic, mode, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 3831 // Restore context register. | 3837 // Restore context register. |
| 3832 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); | 3838 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); |
| 3833 } | 3839 } |
| 3834 | 3840 |
| 3835 | 3841 |
| 3836 void LCodeGen::DoCallFunction(LCallFunction* instr) { | 3842 void LCodeGen::DoCallFunction(LCallFunction* instr) { |
| 3837 ASSERT(ToRegister(instr->function()).is(r1)); | 3843 ASSERT(ToRegister(instr->function()).is(r1)); |
| 3838 ASSERT(ToRegister(instr->result()).is(r0)); | 3844 ASSERT(ToRegister(instr->result()).is(r0)); |
| 3839 | 3845 |
| 3840 int arity = instr->arity(); | 3846 int arity = instr->arity(); |
| 3841 CallFunctionStub stub(arity, NO_CALL_FUNCTION_FLAGS); | 3847 CallFunctionStub stub(arity, NO_CALL_FUNCTION_FLAGS); |
| 3842 CallCode(stub.GetCode(), RelocInfo::CODE_TARGET, instr); | 3848 CallCode(stub.GetCode(), RelocInfo::CODE_TARGET, instr); |
| 3843 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); | 3849 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); |
| 3844 } | 3850 } |
| 3845 | 3851 |
| 3846 | 3852 |
| 3847 void LCodeGen::DoCallGlobal(LCallGlobal* instr) { | 3853 void LCodeGen::DoCallGlobal(LCallGlobal* instr) { |
| 3848 ASSERT(ToRegister(instr->result()).is(r0)); | 3854 ASSERT(ToRegister(instr->result()).is(r0)); |
| 3849 | 3855 |
| 3850 int arity = instr->arity(); | 3856 int arity = instr->arity(); |
| 3851 RelocInfo::Mode mode = RelocInfo::CODE_TARGET_CONTEXT; | 3857 RelocInfo::Mode mode = RelocInfo::CODE_TARGET_CONTEXT; |
| 3852 Handle<Code> ic = | 3858 Handle<Code> ic = |
| 3853 isolate()->stub_cache()->ComputeCallInitialize(arity, mode); | 3859 isolate()->stub_cache()->ComputeCallInitialize(arity, mode); |
| 3854 __ mov(r2, Operand(instr->name())); | 3860 __ mov(r2, Operand(instr->name())); |
| 3855 CallCode(ic, mode, instr); | 3861 CallCode(ic, mode, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 3856 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); | 3862 __ ldr(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); |
| 3857 } | 3863 } |
| 3858 | 3864 |
| 3859 | 3865 |
| 3860 void LCodeGen::DoCallKnownGlobal(LCallKnownGlobal* instr) { | 3866 void LCodeGen::DoCallKnownGlobal(LCallKnownGlobal* instr) { |
| 3861 ASSERT(ToRegister(instr->result()).is(r0)); | 3867 ASSERT(ToRegister(instr->result()).is(r0)); |
| 3862 CallKnownFunction(instr->target(), | 3868 CallKnownFunction(instr->target(), |
| 3863 instr->arity(), | 3869 instr->arity(), |
| 3864 instr, | 3870 instr, |
| 3865 CALL_AS_FUNCTION, | 3871 CALL_AS_FUNCTION, |
| (...skipping 79 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3945 | 3951 |
| 3946 void LCodeGen::DoStoreNamedGeneric(LStoreNamedGeneric* instr) { | 3952 void LCodeGen::DoStoreNamedGeneric(LStoreNamedGeneric* instr) { |
| 3947 ASSERT(ToRegister(instr->object()).is(r1)); | 3953 ASSERT(ToRegister(instr->object()).is(r1)); |
| 3948 ASSERT(ToRegister(instr->value()).is(r0)); | 3954 ASSERT(ToRegister(instr->value()).is(r0)); |
| 3949 | 3955 |
| 3950 // Name is always in r2. | 3956 // Name is always in r2. |
| 3951 __ mov(r2, Operand(instr->name())); | 3957 __ mov(r2, Operand(instr->name())); |
| 3952 Handle<Code> ic = (instr->strict_mode_flag() == kStrictMode) | 3958 Handle<Code> ic = (instr->strict_mode_flag() == kStrictMode) |
| 3953 ? isolate()->builtins()->StoreIC_Initialize_Strict() | 3959 ? isolate()->builtins()->StoreIC_Initialize_Strict() |
| 3954 : isolate()->builtins()->StoreIC_Initialize(); | 3960 : isolate()->builtins()->StoreIC_Initialize(); |
| 3955 CallCode(ic, RelocInfo::CODE_TARGET, instr); | 3961 CallCode(ic, RelocInfo::CODE_TARGET, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 3956 } | 3962 } |
| 3957 | 3963 |
| 3958 | 3964 |
| 3959 void LCodeGen::DeoptIfTaggedButNotSmi(LEnvironment* environment, | 3965 void LCodeGen::DeoptIfTaggedButNotSmi(LEnvironment* environment, |
| 3960 HValue* value, | 3966 HValue* value, |
| 3961 LOperand* operand) { | 3967 LOperand* operand) { |
| 3962 if (value->representation().IsTagged() && !value->type().IsSmi()) { | 3968 if (value->representation().IsTagged() && !value->type().IsSmi()) { |
| 3963 if (operand->IsRegister()) { | 3969 if (operand->IsRegister()) { |
| 3964 __ tst(ToRegister(operand), Operand(kSmiTagMask)); | 3970 __ tst(ToRegister(operand), Operand(kSmiTagMask)); |
| 3965 } else { | 3971 } else { |
| (...skipping 193 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 4159 | 4165 |
| 4160 | 4166 |
| 4161 void LCodeGen::DoStoreKeyedGeneric(LStoreKeyedGeneric* instr) { | 4167 void LCodeGen::DoStoreKeyedGeneric(LStoreKeyedGeneric* instr) { |
| 4162 ASSERT(ToRegister(instr->object()).is(r2)); | 4168 ASSERT(ToRegister(instr->object()).is(r2)); |
| 4163 ASSERT(ToRegister(instr->key()).is(r1)); | 4169 ASSERT(ToRegister(instr->key()).is(r1)); |
| 4164 ASSERT(ToRegister(instr->value()).is(r0)); | 4170 ASSERT(ToRegister(instr->value()).is(r0)); |
| 4165 | 4171 |
| 4166 Handle<Code> ic = (instr->strict_mode_flag() == kStrictMode) | 4172 Handle<Code> ic = (instr->strict_mode_flag() == kStrictMode) |
| 4167 ? isolate()->builtins()->KeyedStoreIC_Initialize_Strict() | 4173 ? isolate()->builtins()->KeyedStoreIC_Initialize_Strict() |
| 4168 : isolate()->builtins()->KeyedStoreIC_Initialize(); | 4174 : isolate()->builtins()->KeyedStoreIC_Initialize(); |
| 4169 CallCode(ic, RelocInfo::CODE_TARGET, instr); | 4175 CallCode(ic, RelocInfo::CODE_TARGET, instr, NEVER_INLINE_TARGET_ADDRESS); |
| 4170 } | 4176 } |
| 4171 | 4177 |
| 4172 | 4178 |
| 4173 void LCodeGen::DoTransitionElementsKind(LTransitionElementsKind* instr) { | 4179 void LCodeGen::DoTransitionElementsKind(LTransitionElementsKind* instr) { |
| 4174 Register object_reg = ToRegister(instr->object()); | 4180 Register object_reg = ToRegister(instr->object()); |
| 4175 Register new_map_reg = ToRegister(instr->new_map_temp()); | 4181 Register new_map_reg = ToRegister(instr->new_map_temp()); |
| 4176 Register scratch = scratch0(); | 4182 Register scratch = scratch0(); |
| 4177 | 4183 |
| 4178 Handle<Map> from_map = instr->original_map(); | 4184 Handle<Map> from_map = instr->original_map(); |
| 4179 Handle<Map> to_map = instr->transitioned_map(); | 4185 Handle<Map> to_map = instr->transitioned_map(); |
| (...skipping 1354 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5534 LEnvironment* env = instr->environment(); | 5540 LEnvironment* env = instr->environment(); |
| 5535 // There is no LLazyBailout instruction for stack-checks. We have to | 5541 // There is no LLazyBailout instruction for stack-checks. We have to |
| 5536 // prepare for lazy deoptimization explicitly here. | 5542 // prepare for lazy deoptimization explicitly here. |
| 5537 if (instr->hydrogen()->is_function_entry()) { | 5543 if (instr->hydrogen()->is_function_entry()) { |
| 5538 // Perform stack overflow check. | 5544 // Perform stack overflow check. |
| 5539 Label done; | 5545 Label done; |
| 5540 __ LoadRoot(ip, Heap::kStackLimitRootIndex); | 5546 __ LoadRoot(ip, Heap::kStackLimitRootIndex); |
| 5541 __ cmp(sp, Operand(ip)); | 5547 __ cmp(sp, Operand(ip)); |
| 5542 __ b(hs, &done); | 5548 __ b(hs, &done); |
| 5543 StackCheckStub stub; | 5549 StackCheckStub stub; |
| 5550 PredictableCodeSizeScope predictable(masm_); |
| 5544 CallCode(stub.GetCode(), RelocInfo::CODE_TARGET, instr); | 5551 CallCode(stub.GetCode(), RelocInfo::CODE_TARGET, instr); |
| 5545 EnsureSpaceForLazyDeopt(); | 5552 EnsureSpaceForLazyDeopt(); |
| 5546 __ bind(&done); | 5553 __ bind(&done); |
| 5547 RegisterEnvironmentForDeoptimization(env, Safepoint::kLazyDeopt); | 5554 RegisterEnvironmentForDeoptimization(env, Safepoint::kLazyDeopt); |
| 5548 safepoints_.RecordLazyDeoptimizationIndex(env->deoptimization_index()); | 5555 safepoints_.RecordLazyDeoptimizationIndex(env->deoptimization_index()); |
| 5549 } else { | 5556 } else { |
| 5550 ASSERT(instr->hydrogen()->is_backwards_branch()); | 5557 ASSERT(instr->hydrogen()->is_backwards_branch()); |
| 5551 // Perform stack overflow check if this goto needs it before jumping. | 5558 // Perform stack overflow check if this goto needs it before jumping. |
| 5552 DeferredStackCheck* deferred_stack_check = | 5559 DeferredStackCheck* deferred_stack_check = |
| 5553 new(zone()) DeferredStackCheck(this, instr); | 5560 new(zone()) DeferredStackCheck(this, instr); |
| (...skipping 119 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5673 __ sub(scratch, result, Operand(index, LSL, kPointerSizeLog2 - kSmiTagSize)); | 5680 __ sub(scratch, result, Operand(index, LSL, kPointerSizeLog2 - kSmiTagSize)); |
| 5674 __ ldr(result, FieldMemOperand(scratch, | 5681 __ ldr(result, FieldMemOperand(scratch, |
| 5675 FixedArray::kHeaderSize - kPointerSize)); | 5682 FixedArray::kHeaderSize - kPointerSize)); |
| 5676 __ bind(&done); | 5683 __ bind(&done); |
| 5677 } | 5684 } |
| 5678 | 5685 |
| 5679 | 5686 |
| 5680 #undef __ | 5687 #undef __ |
| 5681 | 5688 |
| 5682 } } // namespace v8::internal | 5689 } } // namespace v8::internal |
| OLD | NEW |