Index: src/x64/code-stubs-x64.cc |
diff --git a/src/x64/code-stubs-x64.cc b/src/x64/code-stubs-x64.cc |
index 8fbc7dd70d903cf2fa350d80a47dabd5f3f0513c..4c7c99974b50c7a9d6a4b0996a39df52da407bee 100644 |
--- a/src/x64/code-stubs-x64.cc |
+++ b/src/x64/code-stubs-x64.cc |
@@ -322,6 +322,7 @@ static void GenerateFastCloneShallowArrayCommon( |
// |
// rcx: boilerplate literal array. |
ASSERT(mode != FastCloneShallowArrayStub::CLONE_ANY_ELEMENTS); |
+ bool tracking_on = allocation_site_info_mode == TRACK_ALLOCATION_SITE_INFO; |
// All sizes here are multiples of kPointerSize. |
int elements_size = 0; |
@@ -332,10 +333,8 @@ static void GenerateFastCloneShallowArrayCommon( |
} |
int size = JSArray::kSize; |
int allocation_info_start = size; |
- if (allocation_site_info_mode == TRACK_ALLOCATION_SITE_INFO) { |
- size += AllocationSiteInfo::kSize; |
- } |
- size += elements_size; |
+ size += tracking_on ? AllocationSiteInfo::kSize + elements_size |
+ : elements_size; |
// Allocate both the JS array and the elements array in one big |
// allocation. This avoids multiple limit checks. |
@@ -345,7 +344,7 @@ static void GenerateFastCloneShallowArrayCommon( |
} |
__ AllocateInNewSpace(size, rax, rbx, rdx, fail, flags); |
- if (allocation_site_info_mode == TRACK_ALLOCATION_SITE_INFO) { |
+ if (tracking_on) { |
__ LoadRoot(kScratchRegister, Heap::kAllocationSiteInfoMapRootIndex); |
__ movq(FieldOperand(rax, allocation_info_start), kScratchRegister); |
__ movq(FieldOperand(rax, allocation_info_start + kPointerSize), rcx); |
@@ -363,7 +362,7 @@ static void GenerateFastCloneShallowArrayCommon( |
// Get hold of the elements array of the boilerplate and setup the |
// elements pointer in the resulting object. |
__ movq(rcx, FieldOperand(rcx, JSArray::kElementsOffset)); |
- if (allocation_site_info_mode == TRACK_ALLOCATION_SITE_INFO) { |
+ if (tracking_on) { |
__ lea(rdx, Operand(rax, JSArray::kSize + AllocationSiteInfo::kSize)); |
} else { |
__ lea(rdx, Operand(rax, JSArray::kSize)); |
@@ -414,22 +413,14 @@ void FastCloneShallowArrayStub::Generate(MacroAssembler* masm) { |
FastCloneShallowArrayStub::Mode mode = mode_; |
// rcx is boilerplate object. |
Factory* factory = masm->isolate()->factory(); |
- AllocationSiteInfoMode allocation_site_info_mode = |
- DONT_TRACK_ALLOCATION_SITE_INFO; |
- if (mode == CLONE_ANY_ELEMENTS_WITH_ALLOCATION_SITE_INFO) { |
- mode = CLONE_ANY_ELEMENTS; |
- allocation_site_info_mode = TRACK_ALLOCATION_SITE_INFO; |
- } |
- |
if (mode == CLONE_ANY_ELEMENTS) { |
Label double_elements, check_fast_elements; |
__ movq(rbx, FieldOperand(rcx, JSArray::kElementsOffset)); |
__ Cmp(FieldOperand(rbx, HeapObject::kMapOffset), |
factory->fixed_cow_array_map()); |
__ j(not_equal, &check_fast_elements); |
- GenerateFastCloneShallowArrayCommon(masm, 0, |
- COPY_ON_WRITE_ELEMENTS, |
- allocation_site_info_mode, |
+ GenerateFastCloneShallowArrayCommon(masm, 0, COPY_ON_WRITE_ELEMENTS, |
+ allocation_site_info_mode_, |
&slow_case); |
__ ret(3 * kPointerSize); |
@@ -437,9 +428,8 @@ void FastCloneShallowArrayStub::Generate(MacroAssembler* masm) { |
__ Cmp(FieldOperand(rbx, HeapObject::kMapOffset), |
factory->fixed_array_map()); |
__ j(not_equal, &double_elements); |
- GenerateFastCloneShallowArrayCommon(masm, length_, |
- CLONE_ELEMENTS, |
- allocation_site_info_mode, |
+ GenerateFastCloneShallowArrayCommon(masm, length_, CLONE_ELEMENTS, |
+ allocation_site_info_mode_, |
&slow_case); |
__ ret(3 * kPointerSize); |
@@ -471,7 +461,8 @@ void FastCloneShallowArrayStub::Generate(MacroAssembler* masm) { |
} |
GenerateFastCloneShallowArrayCommon(masm, length_, mode, |
- allocation_site_info_mode, &slow_case); |
+ allocation_site_info_mode_, |
+ &slow_case); |
__ ret(3 * kPointerSize); |
__ bind(&slow_case); |