Index: runtime/vm/intermediate_language_arm.cc |
diff --git a/runtime/vm/intermediate_language_arm.cc b/runtime/vm/intermediate_language_arm.cc |
index b4e27f061715c7022d02fcfa15d60492624e1b00..080d9147f6aa5d014b847e21711ae57474ddb6d9 100644 |
--- a/runtime/vm/intermediate_language_arm.cc |
+++ b/runtime/vm/intermediate_language_arm.cc |
@@ -232,7 +232,7 @@ void ClosureCallInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
// deoptimization point in optimized code, after call. |
const intptr_t deopt_id_after = Thread::ToDeoptAfter(deopt_id()); |
if (compiler->is_optimizing()) { |
- compiler->AddDeoptIndexAtCall(deopt_id_after, token_pos()); |
+ compiler->AddDeoptIndexAtCall(deopt_id_after); |
} |
// Add deoptimization continuation point after the call and before the |
// arguments are removed. |
@@ -2842,10 +2842,16 @@ void CatchBlockEntryInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
compiler->assembler()->CodeSize(), |
catch_handler_types_, |
needs_stacktrace()); |
- // Restore the pool pointer. |
- __ RestoreCodePointer(); |
- __ LoadPoolPointer(); |
- |
+ // On lazy deoptimization we patch the optimized code here to enter the |
+ // deoptimization stub. |
+ const intptr_t deopt_id = Thread::ToDeoptAfter(GetDeoptId()); |
+ if (compiler->is_optimizing()) { |
+ compiler->AddDeoptIndexAtCall(deopt_id); |
+ } else { |
+ compiler->AddCurrentDescriptor(RawPcDescriptors::kDeopt, |
+ deopt_id, |
+ TokenPosition::kNoSource); |
+ } |
if (HasParallelMove()) { |
compiler->parallel_move_resolver()->EmitNativeCode(parallel_move()); |
} |