| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 3696 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3707 // Get the value at the masked, scaled index. | 3707 // Get the value at the masked, scaled index. |
| 3708 const int kValueOffset = | 3708 const int kValueOffset = |
| 3709 SeededNumberDictionary::kElementsStartOffset + kPointerSize; | 3709 SeededNumberDictionary::kElementsStartOffset + kPointerSize; |
| 3710 movq(result, FieldOperand(elements, r2, times_pointer_size, kValueOffset)); | 3710 movq(result, FieldOperand(elements, r2, times_pointer_size, kValueOffset)); |
| 3711 } | 3711 } |
| 3712 | 3712 |
| 3713 | 3713 |
| 3714 void MacroAssembler::LoadAllocationTopHelper(Register result, | 3714 void MacroAssembler::LoadAllocationTopHelper(Register result, |
| 3715 Register scratch, | 3715 Register scratch, |
| 3716 AllocationFlags flags) { | 3716 AllocationFlags flags) { |
| 3717 ExternalReference new_space_allocation_top = | 3717 ExternalReference allocation_top = |
| 3718 ExternalReference::new_space_allocation_top_address(isolate()); | 3718 AllocationUtils::GetAllocationTopReference(isolate(), flags); |
| 3719 | 3719 |
| 3720 // Just return if allocation top is already known. | 3720 // Just return if allocation top is already known. |
| 3721 if ((flags & RESULT_CONTAINS_TOP) != 0) { | 3721 if ((flags & RESULT_CONTAINS_TOP) != 0) { |
| 3722 // No use of scratch if allocation top is provided. | 3722 // No use of scratch if allocation top is provided. |
| 3723 ASSERT(!scratch.is_valid()); | 3723 ASSERT(!scratch.is_valid()); |
| 3724 #ifdef DEBUG | 3724 #ifdef DEBUG |
| 3725 // Assert that result actually contains top on entry. | 3725 // Assert that result actually contains top on entry. |
| 3726 Operand top_operand = ExternalOperand(new_space_allocation_top); | 3726 Operand top_operand = ExternalOperand(allocation_top); |
| 3727 cmpq(result, top_operand); | 3727 cmpq(result, top_operand); |
| 3728 Check(equal, "Unexpected allocation top"); | 3728 Check(equal, "Unexpected allocation top"); |
| 3729 #endif | 3729 #endif |
| 3730 return; | 3730 return; |
| 3731 } | 3731 } |
| 3732 | 3732 |
| 3733 // Move address of new object to result. Use scratch register if available, | 3733 // Move address of new object to result. Use scratch register if available, |
| 3734 // and keep address in scratch until call to UpdateAllocationTopHelper. | 3734 // and keep address in scratch until call to UpdateAllocationTopHelper. |
| 3735 if (scratch.is_valid()) { | 3735 if (scratch.is_valid()) { |
| 3736 LoadAddress(scratch, new_space_allocation_top); | 3736 LoadAddress(scratch, allocation_top); |
| 3737 movq(result, Operand(scratch, 0)); | 3737 movq(result, Operand(scratch, 0)); |
| 3738 } else { | 3738 } else { |
| 3739 Load(result, new_space_allocation_top); | 3739 Load(result, allocation_top); |
| 3740 } | 3740 } |
| 3741 } | 3741 } |
| 3742 | 3742 |
| 3743 | 3743 |
| 3744 void MacroAssembler::UpdateAllocationTopHelper(Register result_end, | 3744 void MacroAssembler::UpdateAllocationTopHelper(Register result_end, |
| 3745 Register scratch) { | 3745 Register scratch, |
| 3746 AllocationFlags flags) { |
| 3746 if (emit_debug_code()) { | 3747 if (emit_debug_code()) { |
| 3747 testq(result_end, Immediate(kObjectAlignmentMask)); | 3748 testq(result_end, Immediate(kObjectAlignmentMask)); |
| 3748 Check(zero, "Unaligned allocation in new space"); | 3749 Check(zero, "Unaligned allocation in new space"); |
| 3749 } | 3750 } |
| 3750 | 3751 |
| 3751 ExternalReference new_space_allocation_top = | 3752 ExternalReference allocation_top = |
| 3752 ExternalReference::new_space_allocation_top_address(isolate()); | 3753 AllocationUtils::GetAllocationTopReference(isolate(), flags); |
| 3753 | 3754 |
| 3754 // Update new top. | 3755 // Update new top. |
| 3755 if (scratch.is_valid()) { | 3756 if (scratch.is_valid()) { |
| 3756 // Scratch already contains address of allocation top. | 3757 // Scratch already contains address of allocation top. |
| 3757 movq(Operand(scratch, 0), result_end); | 3758 movq(Operand(scratch, 0), result_end); |
| 3758 } else { | 3759 } else { |
| 3759 Store(new_space_allocation_top, result_end); | 3760 Store(allocation_top, result_end); |
| 3760 } | 3761 } |
| 3761 } | 3762 } |
| 3762 | 3763 |
| 3763 | 3764 |
| 3764 void MacroAssembler::AllocateInNewSpace(int object_size, | 3765 void MacroAssembler::Allocate(int object_size, |
| 3765 Register result, | 3766 Register result, |
| 3766 Register result_end, | 3767 Register result_end, |
| 3767 Register scratch, | 3768 Register scratch, |
| 3768 Label* gc_required, | 3769 Label* gc_required, |
| 3769 AllocationFlags flags) { | 3770 AllocationFlags flags) { |
| 3770 ASSERT((flags & (RESULT_CONTAINS_TOP | SIZE_IN_WORDS)) == 0); | 3771 ASSERT((flags & (RESULT_CONTAINS_TOP | SIZE_IN_WORDS)) == 0); |
| 3771 if (!FLAG_inline_new) { | 3772 if (!FLAG_inline_new) { |
| 3772 if (emit_debug_code()) { | 3773 if (emit_debug_code()) { |
| 3773 // Trash the registers to simulate an allocation failure. | 3774 // Trash the registers to simulate an allocation failure. |
| 3774 movl(result, Immediate(0x7091)); | 3775 movl(result, Immediate(0x7091)); |
| 3775 if (result_end.is_valid()) { | 3776 if (result_end.is_valid()) { |
| 3776 movl(result_end, Immediate(0x7191)); | 3777 movl(result_end, Immediate(0x7191)); |
| 3777 } | 3778 } |
| 3778 if (scratch.is_valid()) { | 3779 if (scratch.is_valid()) { |
| 3779 movl(scratch, Immediate(0x7291)); | 3780 movl(scratch, Immediate(0x7291)); |
| 3780 } | 3781 } |
| 3781 } | 3782 } |
| 3782 jmp(gc_required); | 3783 jmp(gc_required); |
| 3783 return; | 3784 return; |
| 3784 } | 3785 } |
| 3785 ASSERT(!result.is(result_end)); | 3786 ASSERT(!result.is(result_end)); |
| 3786 | 3787 |
| 3787 // Load address of new object into result. | 3788 // Load address of new object into result. |
| 3788 LoadAllocationTopHelper(result, scratch, flags); | 3789 LoadAllocationTopHelper(result, scratch, flags); |
| 3789 | 3790 |
| 3790 // Align the next allocation. Storing the filler map without checking top is | 3791 // Align the next allocation. Storing the filler map without checking top is |
| 3791 // always safe because the limit of the heap is always aligned. | 3792 // always safe because the limit of the heap is always aligned. |
| 3792 if (((flags & DOUBLE_ALIGNMENT) != 0) && FLAG_debug_code) { | 3793 if (((flags & DOUBLE_ALIGNMENT) != 0) && FLAG_debug_code) { |
| 3793 testq(result, Immediate(kDoubleAlignmentMask)); | 3794 testq(result, Immediate(kDoubleAlignmentMask)); |
| 3794 Check(zero, "Allocation is not double aligned"); | 3795 Check(zero, "Allocation is not double aligned"); |
| 3795 } | 3796 } |
| 3796 | 3797 |
| 3797 // Calculate new top and bail out if new space is exhausted. | 3798 // Calculate new top and bail out if new space is exhausted. |
| 3798 ExternalReference new_space_allocation_limit = | 3799 ExternalReference allocation_limit = |
| 3799 ExternalReference::new_space_allocation_limit_address(isolate()); | 3800 AllocationUtils::GetAllocationLimitReference(isolate(), flags); |
| 3800 | 3801 |
| 3801 Register top_reg = result_end.is_valid() ? result_end : result; | 3802 Register top_reg = result_end.is_valid() ? result_end : result; |
| 3802 | 3803 |
| 3803 if (!top_reg.is(result)) { | 3804 if (!top_reg.is(result)) { |
| 3804 movq(top_reg, result); | 3805 movq(top_reg, result); |
| 3805 } | 3806 } |
| 3806 addq(top_reg, Immediate(object_size)); | 3807 addq(top_reg, Immediate(object_size)); |
| 3807 j(carry, gc_required); | 3808 j(carry, gc_required); |
| 3808 Operand limit_operand = ExternalOperand(new_space_allocation_limit); | 3809 Operand limit_operand = ExternalOperand(allocation_limit); |
| 3809 cmpq(top_reg, limit_operand); | 3810 cmpq(top_reg, limit_operand); |
| 3810 j(above, gc_required); | 3811 j(above, gc_required); |
| 3811 | 3812 |
| 3812 // Update allocation top. | 3813 // Update allocation top. |
| 3813 UpdateAllocationTopHelper(top_reg, scratch); | 3814 UpdateAllocationTopHelper(top_reg, scratch, flags); |
| 3814 | 3815 |
| 3815 bool tag_result = (flags & TAG_OBJECT) != 0; | 3816 bool tag_result = (flags & TAG_OBJECT) != 0; |
| 3816 if (top_reg.is(result)) { | 3817 if (top_reg.is(result)) { |
| 3817 if (tag_result) { | 3818 if (tag_result) { |
| 3818 subq(result, Immediate(object_size - kHeapObjectTag)); | 3819 subq(result, Immediate(object_size - kHeapObjectTag)); |
| 3819 } else { | 3820 } else { |
| 3820 subq(result, Immediate(object_size)); | 3821 subq(result, Immediate(object_size)); |
| 3821 } | 3822 } |
| 3822 } else if (tag_result) { | 3823 } else if (tag_result) { |
| 3823 // Tag the result if requested. | 3824 // Tag the result if requested. |
| 3824 ASSERT(kHeapObjectTag == 1); | 3825 ASSERT(kHeapObjectTag == 1); |
| 3825 incq(result); | 3826 incq(result); |
| 3826 } | 3827 } |
| 3827 } | 3828 } |
| 3828 | 3829 |
| 3829 | 3830 |
| 3830 void MacroAssembler::AllocateInNewSpace(int header_size, | 3831 void MacroAssembler::AllocateInNewSpace(int header_size, |
| 3831 ScaleFactor element_size, | 3832 ScaleFactor element_size, |
| 3832 Register element_count, | 3833 Register element_count, |
| 3833 Register result, | 3834 Register result, |
| 3834 Register result_end, | 3835 Register result_end, |
| 3835 Register scratch, | 3836 Register scratch, |
| 3836 Label* gc_required, | 3837 Label* gc_required, |
| 3837 AllocationFlags flags) { | 3838 AllocationFlags flags) { |
| 3838 ASSERT((flags & SIZE_IN_WORDS) == 0); | 3839 ASSERT((flags & SIZE_IN_WORDS) == 0); |
| 3840 ASSERT((flags & PRETENURE_OLD_POINTER_SPACE) == 0); |
| 3839 if (!FLAG_inline_new) { | 3841 if (!FLAG_inline_new) { |
| 3840 if (emit_debug_code()) { | 3842 if (emit_debug_code()) { |
| 3841 // Trash the registers to simulate an allocation failure. | 3843 // Trash the registers to simulate an allocation failure. |
| 3842 movl(result, Immediate(0x7091)); | 3844 movl(result, Immediate(0x7091)); |
| 3843 movl(result_end, Immediate(0x7191)); | 3845 movl(result_end, Immediate(0x7191)); |
| 3844 if (scratch.is_valid()) { | 3846 if (scratch.is_valid()) { |
| 3845 movl(scratch, Immediate(0x7291)); | 3847 movl(scratch, Immediate(0x7291)); |
| 3846 } | 3848 } |
| 3847 // Register element_count is not modified by the function. | 3849 // Register element_count is not modified by the function. |
| 3848 } | 3850 } |
| (...skipping 19 matching lines...) Expand all Loading... |
| 3868 // We assume that element_count*element_size + header_size does not | 3870 // We assume that element_count*element_size + header_size does not |
| 3869 // overflow. | 3871 // overflow. |
| 3870 lea(result_end, Operand(element_count, element_size, header_size)); | 3872 lea(result_end, Operand(element_count, element_size, header_size)); |
| 3871 addq(result_end, result); | 3873 addq(result_end, result); |
| 3872 j(carry, gc_required); | 3874 j(carry, gc_required); |
| 3873 Operand limit_operand = ExternalOperand(new_space_allocation_limit); | 3875 Operand limit_operand = ExternalOperand(new_space_allocation_limit); |
| 3874 cmpq(result_end, limit_operand); | 3876 cmpq(result_end, limit_operand); |
| 3875 j(above, gc_required); | 3877 j(above, gc_required); |
| 3876 | 3878 |
| 3877 // Update allocation top. | 3879 // Update allocation top. |
| 3878 UpdateAllocationTopHelper(result_end, scratch); | 3880 UpdateAllocationTopHelper(result_end, scratch, flags); |
| 3879 | 3881 |
| 3880 // Tag the result if requested. | 3882 // Tag the result if requested. |
| 3881 if ((flags & TAG_OBJECT) != 0) { | 3883 if ((flags & TAG_OBJECT) != 0) { |
| 3882 ASSERT(kHeapObjectTag == 1); | 3884 ASSERT(kHeapObjectTag == 1); |
| 3883 incq(result); | 3885 incq(result); |
| 3884 } | 3886 } |
| 3885 } | 3887 } |
| 3886 | 3888 |
| 3887 | 3889 |
| 3888 void MacroAssembler::AllocateInNewSpace(Register object_size, | 3890 void MacroAssembler::AllocateInNewSpace(Register object_size, |
| 3889 Register result, | 3891 Register result, |
| 3890 Register result_end, | 3892 Register result_end, |
| 3891 Register scratch, | 3893 Register scratch, |
| 3892 Label* gc_required, | 3894 Label* gc_required, |
| 3893 AllocationFlags flags) { | 3895 AllocationFlags flags) { |
| 3894 ASSERT((flags & (RESULT_CONTAINS_TOP | SIZE_IN_WORDS)) == 0); | 3896 ASSERT((flags & (RESULT_CONTAINS_TOP | SIZE_IN_WORDS)) == 0); |
| 3897 ASSERT((flags & PRETENURE_OLD_POINTER_SPACE) == 0); |
| 3895 if (!FLAG_inline_new) { | 3898 if (!FLAG_inline_new) { |
| 3896 if (emit_debug_code()) { | 3899 if (emit_debug_code()) { |
| 3897 // Trash the registers to simulate an allocation failure. | 3900 // Trash the registers to simulate an allocation failure. |
| 3898 movl(result, Immediate(0x7091)); | 3901 movl(result, Immediate(0x7091)); |
| 3899 movl(result_end, Immediate(0x7191)); | 3902 movl(result_end, Immediate(0x7191)); |
| 3900 if (scratch.is_valid()) { | 3903 if (scratch.is_valid()) { |
| 3901 movl(scratch, Immediate(0x7291)); | 3904 movl(scratch, Immediate(0x7291)); |
| 3902 } | 3905 } |
| 3903 // object_size is left unchanged by this function. | 3906 // object_size is left unchanged by this function. |
| 3904 } | 3907 } |
| (...skipping 11 matching lines...) Expand all Loading... |
| 3916 if (!object_size.is(result_end)) { | 3919 if (!object_size.is(result_end)) { |
| 3917 movq(result_end, object_size); | 3920 movq(result_end, object_size); |
| 3918 } | 3921 } |
| 3919 addq(result_end, result); | 3922 addq(result_end, result); |
| 3920 j(carry, gc_required); | 3923 j(carry, gc_required); |
| 3921 Operand limit_operand = ExternalOperand(new_space_allocation_limit); | 3924 Operand limit_operand = ExternalOperand(new_space_allocation_limit); |
| 3922 cmpq(result_end, limit_operand); | 3925 cmpq(result_end, limit_operand); |
| 3923 j(above, gc_required); | 3926 j(above, gc_required); |
| 3924 | 3927 |
| 3925 // Update allocation top. | 3928 // Update allocation top. |
| 3926 UpdateAllocationTopHelper(result_end, scratch); | 3929 UpdateAllocationTopHelper(result_end, scratch, flags); |
| 3927 | 3930 |
| 3928 // Align the next allocation. Storing the filler map without checking top is | 3931 // Align the next allocation. Storing the filler map without checking top is |
| 3929 // always safe because the limit of the heap is always aligned. | 3932 // always safe because the limit of the heap is always aligned. |
| 3930 if (((flags & DOUBLE_ALIGNMENT) != 0) && FLAG_debug_code) { | 3933 if (((flags & DOUBLE_ALIGNMENT) != 0) && FLAG_debug_code) { |
| 3931 testq(result, Immediate(kDoubleAlignmentMask)); | 3934 testq(result, Immediate(kDoubleAlignmentMask)); |
| 3932 Check(zero, "Allocation is not double aligned"); | 3935 Check(zero, "Allocation is not double aligned"); |
| 3933 } | 3936 } |
| 3934 | 3937 |
| 3935 // Tag the result if requested. | 3938 // Tag the result if requested. |
| 3936 if ((flags & TAG_OBJECT) != 0) { | 3939 if ((flags & TAG_OBJECT) != 0) { |
| (...skipping 14 matching lines...) Expand all Loading... |
| 3951 Check(below, "Undo allocation of non allocated memory"); | 3954 Check(below, "Undo allocation of non allocated memory"); |
| 3952 #endif | 3955 #endif |
| 3953 movq(top_operand, object); | 3956 movq(top_operand, object); |
| 3954 } | 3957 } |
| 3955 | 3958 |
| 3956 | 3959 |
| 3957 void MacroAssembler::AllocateHeapNumber(Register result, | 3960 void MacroAssembler::AllocateHeapNumber(Register result, |
| 3958 Register scratch, | 3961 Register scratch, |
| 3959 Label* gc_required) { | 3962 Label* gc_required) { |
| 3960 // Allocate heap number in new space. | 3963 // Allocate heap number in new space. |
| 3961 AllocateInNewSpace(HeapNumber::kSize, | 3964 Allocate(HeapNumber::kSize, result, scratch, no_reg, gc_required, TAG_OBJECT); |
| 3962 result, | |
| 3963 scratch, | |
| 3964 no_reg, | |
| 3965 gc_required, | |
| 3966 TAG_OBJECT); | |
| 3967 | 3965 |
| 3968 // Set the map. | 3966 // Set the map. |
| 3969 LoadRoot(kScratchRegister, Heap::kHeapNumberMapRootIndex); | 3967 LoadRoot(kScratchRegister, Heap::kHeapNumberMapRootIndex); |
| 3970 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); | 3968 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); |
| 3971 } | 3969 } |
| 3972 | 3970 |
| 3973 | 3971 |
| 3974 void MacroAssembler::AllocateTwoByteString(Register result, | 3972 void MacroAssembler::AllocateTwoByteString(Register result, |
| 3975 Register length, | 3973 Register length, |
| 3976 Register scratch1, | 3974 Register scratch1, |
| (...skipping 69 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 4046 movq(FieldOperand(result, String::kHashFieldOffset), | 4044 movq(FieldOperand(result, String::kHashFieldOffset), |
| 4047 Immediate(String::kEmptyHashField)); | 4045 Immediate(String::kEmptyHashField)); |
| 4048 } | 4046 } |
| 4049 | 4047 |
| 4050 | 4048 |
| 4051 void MacroAssembler::AllocateTwoByteConsString(Register result, | 4049 void MacroAssembler::AllocateTwoByteConsString(Register result, |
| 4052 Register scratch1, | 4050 Register scratch1, |
| 4053 Register scratch2, | 4051 Register scratch2, |
| 4054 Label* gc_required) { | 4052 Label* gc_required) { |
| 4055 // Allocate heap number in new space. | 4053 // Allocate heap number in new space. |
| 4056 AllocateInNewSpace(ConsString::kSize, | 4054 Allocate(ConsString::kSize, result, scratch1, scratch2, gc_required, |
| 4057 result, | 4055 TAG_OBJECT); |
| 4058 scratch1, | |
| 4059 scratch2, | |
| 4060 gc_required, | |
| 4061 TAG_OBJECT); | |
| 4062 | 4056 |
| 4063 // Set the map. The other fields are left uninitialized. | 4057 // Set the map. The other fields are left uninitialized. |
| 4064 LoadRoot(kScratchRegister, Heap::kConsStringMapRootIndex); | 4058 LoadRoot(kScratchRegister, Heap::kConsStringMapRootIndex); |
| 4065 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); | 4059 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); |
| 4066 } | 4060 } |
| 4067 | 4061 |
| 4068 | 4062 |
| 4069 void MacroAssembler::AllocateAsciiConsString(Register result, | 4063 void MacroAssembler::AllocateAsciiConsString(Register result, |
| 4070 Register scratch1, | 4064 Register scratch1, |
| 4071 Register scratch2, | 4065 Register scratch2, |
| 4072 Label* gc_required) { | 4066 Label* gc_required) { |
| 4073 // Allocate heap number in new space. | 4067 // Allocate heap number in new space. |
| 4074 AllocateInNewSpace(ConsString::kSize, | 4068 Allocate(ConsString::kSize, result, scratch1, scratch2, gc_required, |
| 4075 result, | 4069 TAG_OBJECT); |
| 4076 scratch1, | |
| 4077 scratch2, | |
| 4078 gc_required, | |
| 4079 TAG_OBJECT); | |
| 4080 | 4070 |
| 4081 // Set the map. The other fields are left uninitialized. | 4071 // Set the map. The other fields are left uninitialized. |
| 4082 LoadRoot(kScratchRegister, Heap::kConsAsciiStringMapRootIndex); | 4072 LoadRoot(kScratchRegister, Heap::kConsAsciiStringMapRootIndex); |
| 4083 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); | 4073 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); |
| 4084 } | 4074 } |
| 4085 | 4075 |
| 4086 | 4076 |
| 4087 void MacroAssembler::AllocateTwoByteSlicedString(Register result, | 4077 void MacroAssembler::AllocateTwoByteSlicedString(Register result, |
| 4088 Register scratch1, | 4078 Register scratch1, |
| 4089 Register scratch2, | 4079 Register scratch2, |
| 4090 Label* gc_required) { | 4080 Label* gc_required) { |
| 4091 // Allocate heap number in new space. | 4081 // Allocate heap number in new space. |
| 4092 AllocateInNewSpace(SlicedString::kSize, | 4082 Allocate(SlicedString::kSize, result, scratch1, scratch2, gc_required, |
| 4093 result, | 4083 TAG_OBJECT); |
| 4094 scratch1, | |
| 4095 scratch2, | |
| 4096 gc_required, | |
| 4097 TAG_OBJECT); | |
| 4098 | 4084 |
| 4099 // Set the map. The other fields are left uninitialized. | 4085 // Set the map. The other fields are left uninitialized. |
| 4100 LoadRoot(kScratchRegister, Heap::kSlicedStringMapRootIndex); | 4086 LoadRoot(kScratchRegister, Heap::kSlicedStringMapRootIndex); |
| 4101 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); | 4087 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); |
| 4102 } | 4088 } |
| 4103 | 4089 |
| 4104 | 4090 |
| 4105 void MacroAssembler::AllocateAsciiSlicedString(Register result, | 4091 void MacroAssembler::AllocateAsciiSlicedString(Register result, |
| 4106 Register scratch1, | 4092 Register scratch1, |
| 4107 Register scratch2, | 4093 Register scratch2, |
| 4108 Label* gc_required) { | 4094 Label* gc_required) { |
| 4109 // Allocate heap number in new space. | 4095 // Allocate heap number in new space. |
| 4110 AllocateInNewSpace(SlicedString::kSize, | 4096 Allocate(SlicedString::kSize, result, scratch1, scratch2, gc_required, |
| 4111 result, | 4097 TAG_OBJECT); |
| 4112 scratch1, | |
| 4113 scratch2, | |
| 4114 gc_required, | |
| 4115 TAG_OBJECT); | |
| 4116 | 4098 |
| 4117 // Set the map. The other fields are left uninitialized. | 4099 // Set the map. The other fields are left uninitialized. |
| 4118 LoadRoot(kScratchRegister, Heap::kSlicedAsciiStringMapRootIndex); | 4100 LoadRoot(kScratchRegister, Heap::kSlicedAsciiStringMapRootIndex); |
| 4119 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); | 4101 movq(FieldOperand(result, HeapObject::kMapOffset), kScratchRegister); |
| 4120 } | 4102 } |
| 4121 | 4103 |
| 4122 | 4104 |
| 4123 // Copy memory, byte-by-byte, from source to destination. Not optimized for | 4105 // Copy memory, byte-by-byte, from source to destination. Not optimized for |
| 4124 // long or aligned copies. The contents of scratch and length are destroyed. | 4106 // long or aligned copies. The contents of scratch and length are destroyed. |
| 4125 // Destination is incremented by length, source, length and scratch are | 4107 // Destination is incremented by length, source, length and scratch are |
| (...skipping 525 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 4651 j(greater, &no_info_available); | 4633 j(greater, &no_info_available); |
| 4652 CompareRoot(MemOperand(scratch_reg, -AllocationSiteInfo::kSize), | 4634 CompareRoot(MemOperand(scratch_reg, -AllocationSiteInfo::kSize), |
| 4653 Heap::kAllocationSiteInfoMapRootIndex); | 4635 Heap::kAllocationSiteInfoMapRootIndex); |
| 4654 bind(&no_info_available); | 4636 bind(&no_info_available); |
| 4655 } | 4637 } |
| 4656 | 4638 |
| 4657 | 4639 |
| 4658 } } // namespace v8::internal | 4640 } } // namespace v8::internal |
| 4659 | 4641 |
| 4660 #endif // V8_TARGET_ARCH_X64 | 4642 #endif // V8_TARGET_ARCH_X64 |
| OLD | NEW |