OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #if V8_TARGET_ARCH_MIPS64 | 5 #if V8_TARGET_ARCH_MIPS64 |
6 | 6 |
7 #include "src/code-stubs.h" | 7 #include "src/code-stubs.h" |
8 #include "src/api-arguments.h" | 8 #include "src/api-arguments.h" |
9 #include "src/bootstrapper.h" | 9 #include "src/bootstrapper.h" |
10 #include "src/codegen.h" | 10 #include "src/codegen.h" |
(...skipping 3792 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3803 | 3803 |
3804 __ bind(&miss); | 3804 __ bind(&miss); |
3805 KeyedLoadIC::GenerateMiss(masm); | 3805 KeyedLoadIC::GenerateMiss(masm); |
3806 | 3806 |
3807 __ bind(&load_smi_map); | 3807 __ bind(&load_smi_map); |
3808 __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); | 3808 __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); |
3809 __ Branch(&compare_map); | 3809 __ Branch(&compare_map); |
3810 } | 3810 } |
3811 | 3811 |
3812 void StoreICTrampolineStub::Generate(MacroAssembler* masm) { | 3812 void StoreICTrampolineStub::Generate(MacroAssembler* masm) { |
3813 __ EmitLoadTypeFeedbackVector(VectorStoreICDescriptor::VectorRegister()); | 3813 __ EmitLoadTypeFeedbackVector(StoreWithVectorDescriptor::VectorRegister()); |
3814 StoreICStub stub(isolate(), state()); | 3814 StoreICStub stub(isolate(), state()); |
3815 stub.GenerateForTrampoline(masm); | 3815 stub.GenerateForTrampoline(masm); |
3816 } | 3816 } |
3817 | 3817 |
3818 void KeyedStoreICTrampolineStub::Generate(MacroAssembler* masm) { | 3818 void KeyedStoreICTrampolineStub::Generate(MacroAssembler* masm) { |
3819 __ EmitLoadTypeFeedbackVector(VectorStoreICDescriptor::VectorRegister()); | 3819 __ EmitLoadTypeFeedbackVector(StoreWithVectorDescriptor::VectorRegister()); |
3820 KeyedStoreICStub stub(isolate(), state()); | 3820 KeyedStoreICStub stub(isolate(), state()); |
3821 stub.GenerateForTrampoline(masm); | 3821 stub.GenerateForTrampoline(masm); |
3822 } | 3822 } |
3823 | 3823 |
3824 void StoreICStub::Generate(MacroAssembler* masm) { GenerateImpl(masm, false); } | 3824 void StoreICStub::Generate(MacroAssembler* masm) { GenerateImpl(masm, false); } |
3825 | 3825 |
3826 void StoreICStub::GenerateForTrampoline(MacroAssembler* masm) { | 3826 void StoreICStub::GenerateForTrampoline(MacroAssembler* masm) { |
3827 GenerateImpl(masm, true); | 3827 GenerateImpl(masm, true); |
3828 } | 3828 } |
3829 | 3829 |
3830 void StoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { | 3830 void StoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { |
3831 Register receiver = VectorStoreICDescriptor::ReceiverRegister(); // a1 | 3831 Register receiver = StoreWithVectorDescriptor::ReceiverRegister(); // a1 |
3832 Register key = VectorStoreICDescriptor::NameRegister(); // a2 | 3832 Register key = StoreWithVectorDescriptor::NameRegister(); // a2 |
3833 Register vector = VectorStoreICDescriptor::VectorRegister(); // a3 | 3833 Register vector = StoreWithVectorDescriptor::VectorRegister(); // a3 |
3834 Register slot = VectorStoreICDescriptor::SlotRegister(); // a4 | 3834 Register slot = StoreWithVectorDescriptor::SlotRegister(); // a4 |
3835 DCHECK(VectorStoreICDescriptor::ValueRegister().is(a0)); // a0 | 3835 DCHECK(StoreWithVectorDescriptor::ValueRegister().is(a0)); // a0 |
3836 Register feedback = a5; | 3836 Register feedback = a5; |
3837 Register receiver_map = a6; | 3837 Register receiver_map = a6; |
3838 Register scratch1 = a7; | 3838 Register scratch1 = a7; |
3839 | 3839 |
3840 __ SmiScale(scratch1, slot, kPointerSizeLog2); | 3840 __ SmiScale(scratch1, slot, kPointerSizeLog2); |
3841 __ Daddu(feedback, vector, Operand(scratch1)); | 3841 __ Daddu(feedback, vector, Operand(scratch1)); |
3842 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); | 3842 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); |
3843 | 3843 |
3844 // Try to quickly handle the monomorphic case without knowing for sure | 3844 // Try to quickly handle the monomorphic case without knowing for sure |
3845 // if we have a weak cell in feedback. We do know it's safe to look | 3845 // if we have a weak cell in feedback. We do know it's safe to look |
(...skipping 90 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3936 | 3936 |
3937 __ bind(&prepare_next); | 3937 __ bind(&prepare_next); |
3938 __ Daddu(pointer_reg, pointer_reg, Operand(kPointerSize * 3)); | 3938 __ Daddu(pointer_reg, pointer_reg, Operand(kPointerSize * 3)); |
3939 __ Branch(&next_loop, lt, pointer_reg, Operand(too_far)); | 3939 __ Branch(&next_loop, lt, pointer_reg, Operand(too_far)); |
3940 | 3940 |
3941 // We exhausted our array of map handler pairs. | 3941 // We exhausted our array of map handler pairs. |
3942 __ Branch(miss); | 3942 __ Branch(miss); |
3943 } | 3943 } |
3944 | 3944 |
3945 void KeyedStoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { | 3945 void KeyedStoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { |
3946 Register receiver = VectorStoreICDescriptor::ReceiverRegister(); // a1 | 3946 Register receiver = StoreWithVectorDescriptor::ReceiverRegister(); // a1 |
3947 Register key = VectorStoreICDescriptor::NameRegister(); // a2 | 3947 Register key = StoreWithVectorDescriptor::NameRegister(); // a2 |
3948 Register vector = VectorStoreICDescriptor::VectorRegister(); // a3 | 3948 Register vector = StoreWithVectorDescriptor::VectorRegister(); // a3 |
3949 Register slot = VectorStoreICDescriptor::SlotRegister(); // a4 | 3949 Register slot = StoreWithVectorDescriptor::SlotRegister(); // a4 |
3950 DCHECK(VectorStoreICDescriptor::ValueRegister().is(a0)); // a0 | 3950 DCHECK(StoreWithVectorDescriptor::ValueRegister().is(a0)); // a0 |
3951 Register feedback = a5; | 3951 Register feedback = a5; |
3952 Register receiver_map = a6; | 3952 Register receiver_map = a6; |
3953 Register scratch1 = a7; | 3953 Register scratch1 = a7; |
3954 | 3954 |
3955 __ SmiScale(scratch1, slot, kPointerSizeLog2); | 3955 __ SmiScale(scratch1, slot, kPointerSizeLog2); |
3956 __ Daddu(feedback, vector, Operand(scratch1)); | 3956 __ Daddu(feedback, vector, Operand(scratch1)); |
3957 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); | 3957 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); |
3958 | 3958 |
3959 // Try to quickly handle the monomorphic case without knowing for sure | 3959 // Try to quickly handle the monomorphic case without knowing for sure |
3960 // if we have a weak cell in feedback. We do know it's safe to look | 3960 // if we have a weak cell in feedback. We do know it's safe to look |
(...skipping 1477 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5438 kStackUnwindSpace, kInvalidStackOffset, | 5438 kStackUnwindSpace, kInvalidStackOffset, |
5439 return_value_operand, NULL); | 5439 return_value_operand, NULL); |
5440 } | 5440 } |
5441 | 5441 |
5442 #undef __ | 5442 #undef __ |
5443 | 5443 |
5444 } // namespace internal | 5444 } // namespace internal |
5445 } // namespace v8 | 5445 } // namespace v8 |
5446 | 5446 |
5447 #endif // V8_TARGET_ARCH_MIPS64 | 5447 #endif // V8_TARGET_ARCH_MIPS64 |
OLD | NEW |