| Index: runtime/vm/flow_graph_compiler_arm64.cc
|
| diff --git a/runtime/vm/flow_graph_compiler_arm64.cc b/runtime/vm/flow_graph_compiler_arm64.cc
|
| index 2634ae2bc7c70e418f6f5afc94bcf7af009faa0e..10ff57f9f8f08a4cc9cf348794d617c993e857fd 100644
|
| --- a/runtime/vm/flow_graph_compiler_arm64.cc
|
| +++ b/runtime/vm/flow_graph_compiler_arm64.cc
|
| @@ -1121,15 +1121,16 @@ void FlowGraphCompiler::CompileGraph() {
|
|
|
| BeginCodeSourceRange();
|
| if (is_optimizing() && !FLAG_precompiled_mode) {
|
| - // Leave enough space for patching in case of lazy deoptimization from
|
| - // deferred code.
|
| + // Leave enough space for patching in case of lazy deoptimization.
|
| for (intptr_t i = 0;
|
| i < CallPattern::kDeoptCallLengthInInstructions;
|
| ++i) {
|
| __ orr(R0, ZR, Operand(R0)); // nop
|
| }
|
| - lazy_deopt_pc_offset_ = assembler()->CodeSize();
|
| - __ BranchPatchable(*StubCode::DeoptimizeLazy_entry());
|
| + lazy_deopt_return_pc_offset_ = assembler()->CodeSize();
|
| + __ BranchPatchable(*StubCode::DeoptimizeLazyFromReturn_entry());
|
| + lazy_deopt_throw_pc_offset_ = assembler()->CodeSize();
|
| + __ BranchPatchable(*StubCode::DeoptimizeLazyFromThrow_entry());
|
| }
|
| EndCodeSourceRange(TokenPosition::kDartCodeEpilogue);
|
| }
|
|
|