| Index: src/compiler/arm64/code-generator-arm64.cc
|
| diff --git a/src/compiler/arm64/code-generator-arm64.cc b/src/compiler/arm64/code-generator-arm64.cc
|
| index ce4336a7b0bfd2df1dc871e64d4493fb73462b85..c3a4f408568716f2ffb0b0fefe3ece17ec6e10cb 100644
|
| --- a/src/compiler/arm64/code-generator-arm64.cc
|
| +++ b/src/compiler/arm64/code-generator-arm64.cc
|
| @@ -545,9 +545,8 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) {
|
| Register value = i.InputRegister(2);
|
| __ Add(index, object, Operand(index, SXTW));
|
| __ Str(value, MemOperand(index));
|
| - SaveFPRegsMode mode = code_->frame()->DidAllocateDoubleRegisters()
|
| - ? kSaveFPRegs
|
| - : kDontSaveFPRegs;
|
| + SaveFPRegsMode mode =
|
| + frame()->DidAllocateDoubleRegisters() ? kSaveFPRegs : kDontSaveFPRegs;
|
| // TODO(dcarney): we shouldn't test write barriers from c calls.
|
| LinkRegisterStatus lr_status = kLRHasNotBeenSaved;
|
| UseScratchRegisterScope scope(masm());
|
| @@ -753,7 +752,7 @@ void CodeGenerator::AssemblePrologue() {
|
| __ PushCalleeSavedRegisters();
|
| frame()->SetRegisterSaveAreaSize(20 * kPointerSize);
|
| } else if (descriptor->IsJSFunctionCall()) {
|
| - CompilationInfo* info = linkage()->info();
|
| + CompilationInfo* info = this->info();
|
| __ SetStackPointer(jssp);
|
| __ Prologue(info->IsCodePreAgingActive());
|
| frame()->SetRegisterSaveAreaSize(
|
| @@ -968,7 +967,7 @@ void CodeGenerator::AddNopForSmiCodeInlining() { __ movz(xzr, 0); }
|
|
|
| void CodeGenerator::EnsureSpaceForLazyDeopt() {
|
| int space_needed = Deoptimizer::patch_size();
|
| - if (!linkage()->info()->IsStub()) {
|
| + if (!info()->IsStub()) {
|
| // Ensure that we have enough space after the previous lazy-bailout
|
| // instruction for patching the code here.
|
| intptr_t current_pc = masm()->pc_offset();
|
|
|