| Index: runtime/vm/flow_graph_compiler_arm64.cc
|
| diff --git a/runtime/vm/flow_graph_compiler_arm64.cc b/runtime/vm/flow_graph_compiler_arm64.cc
|
| index 57c9fd643a9d88c5807ad93135ef027c26029bc7..01d9561ac15be0ae7fe56e3cbf83c034e4dd61db 100644
|
| --- a/runtime/vm/flow_graph_compiler_arm64.cc
|
| +++ b/runtime/vm/flow_graph_compiler_arm64.cc
|
| @@ -1178,20 +1178,19 @@ void FlowGraphCompiler::GenerateRuntimeCall(intptr_t token_pos,
|
| }
|
|
|
|
|
| -void FlowGraphCompiler::EmitEdgeCounter() {
|
| +void FlowGraphCompiler::EmitEdgeCounter(intptr_t edge_id) {
|
| // We do not check for overflow when incrementing the edge counter. The
|
| // function should normally be optimized long before the counter can
|
| // overflow; and though we do not reset the counters when we optimize or
|
| // deoptimize, there is a bound on the number of
|
| // optimization/deoptimization cycles we will attempt.
|
| + ASSERT(!edge_counters_array_.IsNull());
|
| ASSERT(assembler_->constant_pool_allowed());
|
| - const Array& counter = Array::ZoneHandle(zone(), Array::New(1, Heap::kOld));
|
| - counter.SetAt(0, Smi::Handle(zone(), Smi::New(0)));
|
| __ Comment("Edge counter");
|
| - __ LoadUniqueObject(R0, counter);
|
| - __ LoadFieldFromOffset(TMP, R0, Array::element_offset(0));
|
| + __ LoadObject(R0, edge_counters_array_);
|
| + __ LoadFieldFromOffset(TMP, R0, Array::element_offset(edge_id));
|
| __ add(TMP, TMP, Operand(Smi::RawValue(1)));
|
| - __ StoreFieldToOffset(TMP, R0, Array::element_offset(0));
|
| + __ StoreFieldToOffset(TMP, R0, Array::element_offset(edge_id));
|
| }
|
|
|
|
|
|
|