| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #if V8_TARGET_ARCH_MIPS64 | 5 #if V8_TARGET_ARCH_MIPS64 |
| 6 | 6 |
| 7 #include "src/code-stubs.h" | 7 #include "src/code-stubs.h" |
| 8 #include "src/api-arguments.h" | 8 #include "src/api-arguments.h" |
| 9 #include "src/bootstrapper.h" | 9 #include "src/bootstrapper.h" |
| 10 #include "src/codegen.h" | 10 #include "src/codegen.h" |
| (...skipping 3792 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3803 | 3803 |
| 3804 __ bind(&miss); | 3804 __ bind(&miss); |
| 3805 KeyedLoadIC::GenerateMiss(masm); | 3805 KeyedLoadIC::GenerateMiss(masm); |
| 3806 | 3806 |
| 3807 __ bind(&load_smi_map); | 3807 __ bind(&load_smi_map); |
| 3808 __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); | 3808 __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); |
| 3809 __ Branch(&compare_map); | 3809 __ Branch(&compare_map); |
| 3810 } | 3810 } |
| 3811 | 3811 |
| 3812 void StoreICTrampolineStub::Generate(MacroAssembler* masm) { | 3812 void StoreICTrampolineStub::Generate(MacroAssembler* masm) { |
| 3813 __ EmitLoadTypeFeedbackVector(VectorStoreICDescriptor::VectorRegister()); | 3813 __ EmitLoadTypeFeedbackVector(StoreWithVectorDescriptor::VectorRegister()); |
| 3814 StoreICStub stub(isolate(), state()); | 3814 StoreICStub stub(isolate(), state()); |
| 3815 stub.GenerateForTrampoline(masm); | 3815 stub.GenerateForTrampoline(masm); |
| 3816 } | 3816 } |
| 3817 | 3817 |
| 3818 void KeyedStoreICTrampolineStub::Generate(MacroAssembler* masm) { | 3818 void KeyedStoreICTrampolineStub::Generate(MacroAssembler* masm) { |
| 3819 __ EmitLoadTypeFeedbackVector(VectorStoreICDescriptor::VectorRegister()); | 3819 __ EmitLoadTypeFeedbackVector(StoreWithVectorDescriptor::VectorRegister()); |
| 3820 KeyedStoreICStub stub(isolate(), state()); | 3820 KeyedStoreICStub stub(isolate(), state()); |
| 3821 stub.GenerateForTrampoline(masm); | 3821 stub.GenerateForTrampoline(masm); |
| 3822 } | 3822 } |
| 3823 | 3823 |
| 3824 void StoreICStub::Generate(MacroAssembler* masm) { GenerateImpl(masm, false); } | 3824 void StoreICStub::Generate(MacroAssembler* masm) { GenerateImpl(masm, false); } |
| 3825 | 3825 |
| 3826 void StoreICStub::GenerateForTrampoline(MacroAssembler* masm) { | 3826 void StoreICStub::GenerateForTrampoline(MacroAssembler* masm) { |
| 3827 GenerateImpl(masm, true); | 3827 GenerateImpl(masm, true); |
| 3828 } | 3828 } |
| 3829 | 3829 |
| 3830 void StoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { | 3830 void StoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { |
| 3831 Register receiver = VectorStoreICDescriptor::ReceiverRegister(); // a1 | 3831 Register receiver = StoreWithVectorDescriptor::ReceiverRegister(); // a1 |
| 3832 Register key = VectorStoreICDescriptor::NameRegister(); // a2 | 3832 Register key = StoreWithVectorDescriptor::NameRegister(); // a2 |
| 3833 Register vector = VectorStoreICDescriptor::VectorRegister(); // a3 | 3833 Register vector = StoreWithVectorDescriptor::VectorRegister(); // a3 |
| 3834 Register slot = VectorStoreICDescriptor::SlotRegister(); // a4 | 3834 Register slot = StoreWithVectorDescriptor::SlotRegister(); // a4 |
| 3835 DCHECK(VectorStoreICDescriptor::ValueRegister().is(a0)); // a0 | 3835 DCHECK(StoreWithVectorDescriptor::ValueRegister().is(a0)); // a0 |
| 3836 Register feedback = a5; | 3836 Register feedback = a5; |
| 3837 Register receiver_map = a6; | 3837 Register receiver_map = a6; |
| 3838 Register scratch1 = a7; | 3838 Register scratch1 = a7; |
| 3839 | 3839 |
| 3840 __ SmiScale(scratch1, slot, kPointerSizeLog2); | 3840 __ SmiScale(scratch1, slot, kPointerSizeLog2); |
| 3841 __ Daddu(feedback, vector, Operand(scratch1)); | 3841 __ Daddu(feedback, vector, Operand(scratch1)); |
| 3842 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); | 3842 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); |
| 3843 | 3843 |
| 3844 // Try to quickly handle the monomorphic case without knowing for sure | 3844 // Try to quickly handle the monomorphic case without knowing for sure |
| 3845 // if we have a weak cell in feedback. We do know it's safe to look | 3845 // if we have a weak cell in feedback. We do know it's safe to look |
| (...skipping 90 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3936 | 3936 |
| 3937 __ bind(&prepare_next); | 3937 __ bind(&prepare_next); |
| 3938 __ Daddu(pointer_reg, pointer_reg, Operand(kPointerSize * 3)); | 3938 __ Daddu(pointer_reg, pointer_reg, Operand(kPointerSize * 3)); |
| 3939 __ Branch(&next_loop, lt, pointer_reg, Operand(too_far)); | 3939 __ Branch(&next_loop, lt, pointer_reg, Operand(too_far)); |
| 3940 | 3940 |
| 3941 // We exhausted our array of map handler pairs. | 3941 // We exhausted our array of map handler pairs. |
| 3942 __ Branch(miss); | 3942 __ Branch(miss); |
| 3943 } | 3943 } |
| 3944 | 3944 |
| 3945 void KeyedStoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { | 3945 void KeyedStoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { |
| 3946 Register receiver = VectorStoreICDescriptor::ReceiverRegister(); // a1 | 3946 Register receiver = StoreWithVectorDescriptor::ReceiverRegister(); // a1 |
| 3947 Register key = VectorStoreICDescriptor::NameRegister(); // a2 | 3947 Register key = StoreWithVectorDescriptor::NameRegister(); // a2 |
| 3948 Register vector = VectorStoreICDescriptor::VectorRegister(); // a3 | 3948 Register vector = StoreWithVectorDescriptor::VectorRegister(); // a3 |
| 3949 Register slot = VectorStoreICDescriptor::SlotRegister(); // a4 | 3949 Register slot = StoreWithVectorDescriptor::SlotRegister(); // a4 |
| 3950 DCHECK(VectorStoreICDescriptor::ValueRegister().is(a0)); // a0 | 3950 DCHECK(StoreWithVectorDescriptor::ValueRegister().is(a0)); // a0 |
| 3951 Register feedback = a5; | 3951 Register feedback = a5; |
| 3952 Register receiver_map = a6; | 3952 Register receiver_map = a6; |
| 3953 Register scratch1 = a7; | 3953 Register scratch1 = a7; |
| 3954 | 3954 |
| 3955 __ SmiScale(scratch1, slot, kPointerSizeLog2); | 3955 __ SmiScale(scratch1, slot, kPointerSizeLog2); |
| 3956 __ Daddu(feedback, vector, Operand(scratch1)); | 3956 __ Daddu(feedback, vector, Operand(scratch1)); |
| 3957 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); | 3957 __ ld(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); |
| 3958 | 3958 |
| 3959 // Try to quickly handle the monomorphic case without knowing for sure | 3959 // Try to quickly handle the monomorphic case without knowing for sure |
| 3960 // if we have a weak cell in feedback. We do know it's safe to look | 3960 // if we have a weak cell in feedback. We do know it's safe to look |
| (...skipping 1477 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5438 kStackUnwindSpace, kInvalidStackOffset, | 5438 kStackUnwindSpace, kInvalidStackOffset, |
| 5439 return_value_operand, NULL); | 5439 return_value_operand, NULL); |
| 5440 } | 5440 } |
| 5441 | 5441 |
| 5442 #undef __ | 5442 #undef __ |
| 5443 | 5443 |
| 5444 } // namespace internal | 5444 } // namespace internal |
| 5445 } // namespace v8 | 5445 } // namespace v8 |
| 5446 | 5446 |
| 5447 #endif // V8_TARGET_ARCH_MIPS64 | 5447 #endif // V8_TARGET_ARCH_MIPS64 |
| OLD | NEW |