| Index: runtime/vm/stub_code_x64.cc
|
| diff --git a/runtime/vm/stub_code_x64.cc b/runtime/vm/stub_code_x64.cc
|
| index 157f7662abf40bd137e2165349220b79b207166b..f527a8a17c7260ec74f9698a332016a3675e2acd 100644
|
| --- a/runtime/vm/stub_code_x64.cc
|
| +++ b/runtime/vm/stub_code_x64.cc
|
| @@ -646,9 +646,8 @@ void StubCode::GenerateAllocateArrayStub(Assembler* assembler) {
|
| __ andq(RDI, Immediate(-kObjectAlignment));
|
|
|
| const intptr_t cid = kArrayCid;
|
| - Heap::Space space = Heap::kNew;
|
| - __ movq(R13, Address(THR, Thread::heap_offset()));
|
| - __ movq(RAX, Address(R13, Heap::TopOffset(space)));
|
| + NOT_IN_PRODUCT(Heap::Space space = Heap::kNew);
|
| + __ movq(RAX, Address(THR, Thread::top_offset()));
|
|
|
| // RDI: allocation size.
|
| __ movq(RCX, RAX);
|
| @@ -659,13 +658,12 @@ void StubCode::GenerateAllocateArrayStub(Assembler* assembler) {
|
| // RAX: potential new object start.
|
| // RCX: potential next object start.
|
| // RDI: allocation size.
|
| - // R13: heap.
|
| - __ cmpq(RCX, Address(R13, Heap::EndOffset(space)));
|
| + __ cmpq(RCX, Address(THR, Thread::end_offset()));
|
| __ j(ABOVE_EQUAL, &slow_case);
|
|
|
| // Successfully allocated the object(s), now update top to point to
|
| // next object start and initialize the object.
|
| - __ movq(Address(R13, Heap::TopOffset(space)), RCX);
|
| + __ movq(Address(THR, Thread::top_offset()), RCX);
|
| __ addq(RAX, Immediate(kHeapObjectTag));
|
| NOT_IN_PRODUCT(__ UpdateAllocationStatsWithSize(cid, RDI, space));
|
| // Initialize the tags.
|
| @@ -893,16 +891,14 @@ void StubCode::GenerateAllocateContextStub(Assembler* assembler) {
|
| // Now allocate the object.
|
| // R10: number of context variables.
|
| const intptr_t cid = kContextCid;
|
| - Heap::Space space = Heap::kNew;
|
| - __ movq(RCX, Address(THR, Thread::heap_offset()));
|
| - __ movq(RAX, Address(RCX, Heap::TopOffset(space)));
|
| + NOT_IN_PRODUCT(Heap::Space space = Heap::kNew);
|
| + __ movq(RAX, Address(THR, Thread::top_offset()));
|
| __ addq(R13, RAX);
|
| // Check if the allocation fits into the remaining space.
|
| // RAX: potential new object.
|
| // R13: potential next object start.
|
| // R10: number of context variables.
|
| - // RCX: heap.
|
| - __ cmpq(R13, Address(RCX, Heap::EndOffset(space)));
|
| + __ cmpq(R13, Address(THR, Thread::end_offset()));
|
| if (FLAG_use_slow_path) {
|
| __ jmp(&slow_case);
|
| } else {
|
| @@ -914,8 +910,7 @@ void StubCode::GenerateAllocateContextStub(Assembler* assembler) {
|
| // RAX: new object.
|
| // R13: next object start.
|
| // R10: number of context variables.
|
| - // RCX: heap.
|
| - __ movq(Address(RCX, Heap::TopOffset(space)), R13);
|
| + __ movq(Address(THR, Thread::top_offset()), R13);
|
| // R13: Size of allocation in bytes.
|
| __ subq(R13, RAX);
|
| __ addq(RAX, Immediate(kHeapObjectTag));
|
| @@ -1092,21 +1087,19 @@ void StubCode::GenerateAllocationStubForClass(Assembler* assembler,
|
| // Allocate the object and update top to point to
|
| // next object start and initialize the allocated object.
|
| // RDX: instantiated type arguments (if is_cls_parameterized).
|
| - Heap::Space space = Heap::kNew;
|
| - __ movq(RCX, Address(THR, Thread::heap_offset()));
|
| - __ movq(RAX, Address(RCX, Heap::TopOffset(space)));
|
| + NOT_IN_PRODUCT(Heap::Space space = Heap::kNew);
|
| + __ movq(RAX, Address(THR, Thread::top_offset()));
|
| __ leaq(RBX, Address(RAX, instance_size));
|
| // Check if the allocation fits into the remaining space.
|
| // RAX: potential new object start.
|
| // RBX: potential next object start.
|
| - // RCX: heap.
|
| - __ cmpq(RBX, Address(RCX, Heap::EndOffset(space)));
|
| + __ cmpq(RBX, Address(THR, Thread::end_offset()));
|
| if (FLAG_use_slow_path) {
|
| __ jmp(&slow_case);
|
| } else {
|
| __ j(ABOVE_EQUAL, &slow_case);
|
| }
|
| - __ movq(Address(RCX, Heap::TopOffset(space)), RBX);
|
| + __ movq(Address(THR, Thread::top_offset()), RBX);
|
| NOT_IN_PRODUCT(__ UpdateAllocationStats(cls.id(), space));
|
|
|
| // RAX: new object start (untagged).
|
|
|