OLD | NEW |
1 // Copyright 2014 the V8 project authors. All rights reserved. | 1 // Copyright 2014 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #if V8_TARGET_ARCH_PPC | 5 #if V8_TARGET_ARCH_PPC |
6 | 6 |
7 #include "src/code-stubs.h" | 7 #include "src/code-stubs.h" |
8 #include "src/api-arguments.h" | 8 #include "src/api-arguments.h" |
9 #include "src/base/bits.h" | 9 #include "src/base/bits.h" |
10 #include "src/bootstrapper.h" | 10 #include "src/bootstrapper.h" |
(...skipping 3720 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3731 | 3731 |
3732 __ bind(&miss); | 3732 __ bind(&miss); |
3733 KeyedLoadIC::GenerateMiss(masm); | 3733 KeyedLoadIC::GenerateMiss(masm); |
3734 | 3734 |
3735 __ bind(&load_smi_map); | 3735 __ bind(&load_smi_map); |
3736 __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); | 3736 __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); |
3737 __ b(&compare_map); | 3737 __ b(&compare_map); |
3738 } | 3738 } |
3739 | 3739 |
3740 void StoreICTrampolineStub::Generate(MacroAssembler* masm) { | 3740 void StoreICTrampolineStub::Generate(MacroAssembler* masm) { |
3741 __ EmitLoadTypeFeedbackVector(VectorStoreICDescriptor::VectorRegister()); | 3741 __ EmitLoadTypeFeedbackVector(StoreWithVectorDescriptor::VectorRegister()); |
3742 StoreICStub stub(isolate(), state()); | 3742 StoreICStub stub(isolate(), state()); |
3743 stub.GenerateForTrampoline(masm); | 3743 stub.GenerateForTrampoline(masm); |
3744 } | 3744 } |
3745 | 3745 |
3746 void KeyedStoreICTrampolineStub::Generate(MacroAssembler* masm) { | 3746 void KeyedStoreICTrampolineStub::Generate(MacroAssembler* masm) { |
3747 __ EmitLoadTypeFeedbackVector(VectorStoreICDescriptor::VectorRegister()); | 3747 __ EmitLoadTypeFeedbackVector(StoreWithVectorDescriptor::VectorRegister()); |
3748 KeyedStoreICStub stub(isolate(), state()); | 3748 KeyedStoreICStub stub(isolate(), state()); |
3749 stub.GenerateForTrampoline(masm); | 3749 stub.GenerateForTrampoline(masm); |
3750 } | 3750 } |
3751 | 3751 |
3752 void StoreICStub::Generate(MacroAssembler* masm) { GenerateImpl(masm, false); } | 3752 void StoreICStub::Generate(MacroAssembler* masm) { GenerateImpl(masm, false); } |
3753 | 3753 |
3754 void StoreICStub::GenerateForTrampoline(MacroAssembler* masm) { | 3754 void StoreICStub::GenerateForTrampoline(MacroAssembler* masm) { |
3755 GenerateImpl(masm, true); | 3755 GenerateImpl(masm, true); |
3756 } | 3756 } |
3757 | 3757 |
3758 void StoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { | 3758 void StoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { |
3759 Register receiver = VectorStoreICDescriptor::ReceiverRegister(); // r4 | 3759 Register receiver = StoreWithVectorDescriptor::ReceiverRegister(); // r4 |
3760 Register key = VectorStoreICDescriptor::NameRegister(); // r5 | 3760 Register key = StoreWithVectorDescriptor::NameRegister(); // r5 |
3761 Register vector = VectorStoreICDescriptor::VectorRegister(); // r6 | 3761 Register vector = StoreWithVectorDescriptor::VectorRegister(); // r6 |
3762 Register slot = VectorStoreICDescriptor::SlotRegister(); // r7 | 3762 Register slot = StoreWithVectorDescriptor::SlotRegister(); // r7 |
3763 DCHECK(VectorStoreICDescriptor::ValueRegister().is(r3)); // r3 | 3763 DCHECK(StoreWithVectorDescriptor::ValueRegister().is(r3)); // r3 |
3764 Register feedback = r8; | 3764 Register feedback = r8; |
3765 Register receiver_map = r9; | 3765 Register receiver_map = r9; |
3766 Register scratch1 = r10; | 3766 Register scratch1 = r10; |
3767 | 3767 |
3768 __ SmiToPtrArrayOffset(r0, slot); | 3768 __ SmiToPtrArrayOffset(r0, slot); |
3769 __ add(feedback, vector, r0); | 3769 __ add(feedback, vector, r0); |
3770 __ LoadP(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); | 3770 __ LoadP(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); |
3771 | 3771 |
3772 // Try to quickly handle the monomorphic case without knowing for sure | 3772 // Try to quickly handle the monomorphic case without knowing for sure |
3773 // if we have a weak cell in feedback. We do know it's safe to look | 3773 // if we have a weak cell in feedback. We do know it's safe to look |
(...skipping 94 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3868 __ bind(&prepare_next); | 3868 __ bind(&prepare_next); |
3869 __ addi(pointer_reg, pointer_reg, Operand(kPointerSize * 3)); | 3869 __ addi(pointer_reg, pointer_reg, Operand(kPointerSize * 3)); |
3870 __ cmpl(pointer_reg, too_far); | 3870 __ cmpl(pointer_reg, too_far); |
3871 __ blt(&next_loop); | 3871 __ blt(&next_loop); |
3872 | 3872 |
3873 // We exhausted our array of map handler pairs. | 3873 // We exhausted our array of map handler pairs. |
3874 __ b(miss); | 3874 __ b(miss); |
3875 } | 3875 } |
3876 | 3876 |
3877 void KeyedStoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { | 3877 void KeyedStoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { |
3878 Register receiver = VectorStoreICDescriptor::ReceiverRegister(); // r4 | 3878 Register receiver = StoreWithVectorDescriptor::ReceiverRegister(); // r4 |
3879 Register key = VectorStoreICDescriptor::NameRegister(); // r5 | 3879 Register key = StoreWithVectorDescriptor::NameRegister(); // r5 |
3880 Register vector = VectorStoreICDescriptor::VectorRegister(); // r6 | 3880 Register vector = StoreWithVectorDescriptor::VectorRegister(); // r6 |
3881 Register slot = VectorStoreICDescriptor::SlotRegister(); // r7 | 3881 Register slot = StoreWithVectorDescriptor::SlotRegister(); // r7 |
3882 DCHECK(VectorStoreICDescriptor::ValueRegister().is(r3)); // r3 | 3882 DCHECK(StoreWithVectorDescriptor::ValueRegister().is(r3)); // r3 |
3883 Register feedback = r8; | 3883 Register feedback = r8; |
3884 Register receiver_map = r9; | 3884 Register receiver_map = r9; |
3885 Register scratch1 = r10; | 3885 Register scratch1 = r10; |
3886 | 3886 |
3887 __ SmiToPtrArrayOffset(r0, slot); | 3887 __ SmiToPtrArrayOffset(r0, slot); |
3888 __ add(feedback, vector, r0); | 3888 __ add(feedback, vector, r0); |
3889 __ LoadP(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); | 3889 __ LoadP(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); |
3890 | 3890 |
3891 // Try to quickly handle the monomorphic case without knowing for sure | 3891 // Try to quickly handle the monomorphic case without knowing for sure |
3892 // if we have a weak cell in feedback. We do know it's safe to look | 3892 // if we have a weak cell in feedback. We do know it's safe to look |
(...skipping 1584 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5477 fp, (PropertyCallbackArguments::kReturnValueOffset + 3) * kPointerSize); | 5477 fp, (PropertyCallbackArguments::kReturnValueOffset + 3) * kPointerSize); |
5478 CallApiFunctionAndReturn(masm, api_function_address, thunk_ref, | 5478 CallApiFunctionAndReturn(masm, api_function_address, thunk_ref, |
5479 kStackUnwindSpace, NULL, return_value_operand, NULL); | 5479 kStackUnwindSpace, NULL, return_value_operand, NULL); |
5480 } | 5480 } |
5481 | 5481 |
5482 #undef __ | 5482 #undef __ |
5483 } // namespace internal | 5483 } // namespace internal |
5484 } // namespace v8 | 5484 } // namespace v8 |
5485 | 5485 |
5486 #endif // V8_TARGET_ARCH_PPC | 5486 #endif // V8_TARGET_ARCH_PPC |
OLD | NEW |