Index: runtime/vm/stub_code_x64.cc |
diff --git a/runtime/vm/stub_code_x64.cc b/runtime/vm/stub_code_x64.cc |
index 157f7662abf40bd137e2165349220b79b207166b..f527a8a17c7260ec74f9698a332016a3675e2acd 100644 |
--- a/runtime/vm/stub_code_x64.cc |
+++ b/runtime/vm/stub_code_x64.cc |
@@ -646,9 +646,8 @@ void StubCode::GenerateAllocateArrayStub(Assembler* assembler) { |
__ andq(RDI, Immediate(-kObjectAlignment)); |
const intptr_t cid = kArrayCid; |
- Heap::Space space = Heap::kNew; |
- __ movq(R13, Address(THR, Thread::heap_offset())); |
- __ movq(RAX, Address(R13, Heap::TopOffset(space))); |
+ NOT_IN_PRODUCT(Heap::Space space = Heap::kNew); |
+ __ movq(RAX, Address(THR, Thread::top_offset())); |
// RDI: allocation size. |
__ movq(RCX, RAX); |
@@ -659,13 +658,12 @@ void StubCode::GenerateAllocateArrayStub(Assembler* assembler) { |
// RAX: potential new object start. |
// RCX: potential next object start. |
// RDI: allocation size. |
- // R13: heap. |
- __ cmpq(RCX, Address(R13, Heap::EndOffset(space))); |
+ __ cmpq(RCX, Address(THR, Thread::end_offset())); |
__ j(ABOVE_EQUAL, &slow_case); |
// Successfully allocated the object(s), now update top to point to |
// next object start and initialize the object. |
- __ movq(Address(R13, Heap::TopOffset(space)), RCX); |
+ __ movq(Address(THR, Thread::top_offset()), RCX); |
__ addq(RAX, Immediate(kHeapObjectTag)); |
NOT_IN_PRODUCT(__ UpdateAllocationStatsWithSize(cid, RDI, space)); |
// Initialize the tags. |
@@ -893,16 +891,14 @@ void StubCode::GenerateAllocateContextStub(Assembler* assembler) { |
// Now allocate the object. |
// R10: number of context variables. |
const intptr_t cid = kContextCid; |
- Heap::Space space = Heap::kNew; |
- __ movq(RCX, Address(THR, Thread::heap_offset())); |
- __ movq(RAX, Address(RCX, Heap::TopOffset(space))); |
+ NOT_IN_PRODUCT(Heap::Space space = Heap::kNew); |
+ __ movq(RAX, Address(THR, Thread::top_offset())); |
__ addq(R13, RAX); |
// Check if the allocation fits into the remaining space. |
// RAX: potential new object. |
// R13: potential next object start. |
// R10: number of context variables. |
- // RCX: heap. |
- __ cmpq(R13, Address(RCX, Heap::EndOffset(space))); |
+ __ cmpq(R13, Address(THR, Thread::end_offset())); |
if (FLAG_use_slow_path) { |
__ jmp(&slow_case); |
} else { |
@@ -914,8 +910,7 @@ void StubCode::GenerateAllocateContextStub(Assembler* assembler) { |
// RAX: new object. |
// R13: next object start. |
// R10: number of context variables. |
- // RCX: heap. |
- __ movq(Address(RCX, Heap::TopOffset(space)), R13); |
+ __ movq(Address(THR, Thread::top_offset()), R13); |
// R13: Size of allocation in bytes. |
__ subq(R13, RAX); |
__ addq(RAX, Immediate(kHeapObjectTag)); |
@@ -1092,21 +1087,19 @@ void StubCode::GenerateAllocationStubForClass(Assembler* assembler, |
// Allocate the object and update top to point to |
// next object start and initialize the allocated object. |
// RDX: instantiated type arguments (if is_cls_parameterized). |
- Heap::Space space = Heap::kNew; |
- __ movq(RCX, Address(THR, Thread::heap_offset())); |
- __ movq(RAX, Address(RCX, Heap::TopOffset(space))); |
+ NOT_IN_PRODUCT(Heap::Space space = Heap::kNew); |
+ __ movq(RAX, Address(THR, Thread::top_offset())); |
__ leaq(RBX, Address(RAX, instance_size)); |
// Check if the allocation fits into the remaining space. |
// RAX: potential new object start. |
// RBX: potential next object start. |
- // RCX: heap. |
- __ cmpq(RBX, Address(RCX, Heap::EndOffset(space))); |
+ __ cmpq(RBX, Address(THR, Thread::end_offset())); |
if (FLAG_use_slow_path) { |
__ jmp(&slow_case); |
} else { |
__ j(ABOVE_EQUAL, &slow_case); |
} |
- __ movq(Address(RCX, Heap::TopOffset(space)), RBX); |
+ __ movq(Address(THR, Thread::top_offset()), RBX); |
NOT_IN_PRODUCT(__ UpdateAllocationStats(cls.id(), space)); |
// RAX: new object start (untagged). |