| Index: runtime/vm/intermediate_language_arm.cc
|
| diff --git a/runtime/vm/intermediate_language_arm.cc b/runtime/vm/intermediate_language_arm.cc
|
| index b4e27f061715c7022d02fcfa15d60492624e1b00..080d9147f6aa5d014b847e21711ae57474ddb6d9 100644
|
| --- a/runtime/vm/intermediate_language_arm.cc
|
| +++ b/runtime/vm/intermediate_language_arm.cc
|
| @@ -232,7 +232,7 @@ void ClosureCallInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| // deoptimization point in optimized code, after call.
|
| const intptr_t deopt_id_after = Thread::ToDeoptAfter(deopt_id());
|
| if (compiler->is_optimizing()) {
|
| - compiler->AddDeoptIndexAtCall(deopt_id_after, token_pos());
|
| + compiler->AddDeoptIndexAtCall(deopt_id_after);
|
| }
|
| // Add deoptimization continuation point after the call and before the
|
| // arguments are removed.
|
| @@ -2842,10 +2842,16 @@ void CatchBlockEntryInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| compiler->assembler()->CodeSize(),
|
| catch_handler_types_,
|
| needs_stacktrace());
|
| - // Restore the pool pointer.
|
| - __ RestoreCodePointer();
|
| - __ LoadPoolPointer();
|
| -
|
| + // On lazy deoptimization we patch the optimized code here to enter the
|
| + // deoptimization stub.
|
| + const intptr_t deopt_id = Thread::ToDeoptAfter(GetDeoptId());
|
| + if (compiler->is_optimizing()) {
|
| + compiler->AddDeoptIndexAtCall(deopt_id);
|
| + } else {
|
| + compiler->AddCurrentDescriptor(RawPcDescriptors::kDeopt,
|
| + deopt_id,
|
| + TokenPosition::kNoSource);
|
| + }
|
| if (HasParallelMove()) {
|
| compiler->parallel_move_resolver()->EmitNativeCode(parallel_move());
|
| }
|
|
|