| Index: runtime/vm/constant_propagator.cc
 | 
| ===================================================================
 | 
| --- runtime/vm/constant_propagator.cc	(revision 42566)
 | 
| +++ runtime/vm/constant_propagator.cc	(working copy)
 | 
| @@ -2,7504 +2,27 @@
 | 
|  // for details. All rights reserved. Use of this source code is governed by a
 | 
|  // BSD-style license that can be found in the LICENSE file.
 | 
|  
 | 
| -#include "vm/flow_graph_optimizer.h"
 | 
| +#include "vm/constant_propagator.h"
 | 
|  
 | 
|  #include "vm/bit_vector.h"
 | 
| -#include "vm/cha.h"
 | 
| -#include "vm/cpu.h"
 | 
| -#include "vm/dart_entry.h"
 | 
| -#include "vm/exceptions.h"
 | 
|  #include "vm/flow_graph_builder.h"
 | 
|  #include "vm/flow_graph_compiler.h"
 | 
|  #include "vm/flow_graph_range_analysis.h"
 | 
| -#include "vm/hash_map.h"
 | 
|  #include "vm/il_printer.h"
 | 
|  #include "vm/intermediate_language.h"
 | 
| -#include "vm/object_store.h"
 | 
|  #include "vm/parser.h"
 | 
| -#include "vm/resolver.h"
 | 
| -#include "vm/scopes.h"
 | 
| -#include "vm/stack_frame.h"
 | 
|  #include "vm/symbols.h"
 | 
|  
 | 
|  namespace dart {
 | 
|  
 | 
| -DEFINE_FLAG(int, getter_setter_ratio, 13,
 | 
| -    "Ratio of getter/setter usage used for double field unboxing heuristics");
 | 
| -DEFINE_FLAG(bool, load_cse, true, "Use redundant load elimination.");
 | 
| -DEFINE_FLAG(bool, dead_store_elimination, true, "Eliminate dead stores");
 | 
| -DEFINE_FLAG(int, max_polymorphic_checks, 4,
 | 
| -    "Maximum number of polymorphic check, otherwise it is megamorphic.");
 | 
| -DEFINE_FLAG(int, max_equality_polymorphic_checks, 32,
 | 
| -    "Maximum number of polymorphic checks in equality operator,"
 | 
| -    " otherwise use megamorphic dispatch.");
 | 
|  DEFINE_FLAG(bool, remove_redundant_phis, true, "Remove redundant phis.");
 | 
|  DEFINE_FLAG(bool, trace_constant_propagation, false,
 | 
|      "Print constant propagation and useless code elimination.");
 | 
| -DEFINE_FLAG(bool, trace_load_optimization, false,
 | 
| -    "Print live sets for load optimization pass.");
 | 
| -DEFINE_FLAG(bool, trace_optimization, false, "Print optimization details.");
 | 
| -DEFINE_FLAG(bool, truncating_left_shift, true,
 | 
| -    "Optimize left shift to truncate if possible");
 | 
| -DEFINE_FLAG(bool, use_cha, true, "Use class hierarchy analysis.");
 | 
| -#if defined(TARGET_ARCH_ARM) || defined(TARGET_ARCH_IA32)
 | 
| -DEFINE_FLAG(bool, trace_smi_widening, false, "Trace Smi->Int32 widening pass.");
 | 
| -#endif
 | 
| -DECLARE_FLAG(bool, enable_type_checks);
 | 
| -DECLARE_FLAG(bool, source_lines);
 | 
| -DECLARE_FLAG(bool, trace_type_check_elimination);
 | 
| -DECLARE_FLAG(bool, warn_on_javascript_compatibility);
 | 
|  
 | 
|  // Quick access to the locally defined isolate() method.
 | 
|  #define I (isolate())
 | 
|  
 | 
| -static bool ShouldInlineSimd() {
 | 
| -  return FlowGraphCompiler::SupportsUnboxedSimd128();
 | 
| -}
 | 
|  
 | 
| -
 | 
| -static bool CanUnboxDouble() {
 | 
| -  return FlowGraphCompiler::SupportsUnboxedDoubles();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool ShouldInlineInt64ArrayOps() {
 | 
| -#if defined(TARGET_ARCH_X64)
 | 
| -  return true;
 | 
| -#endif
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -static bool CanConvertUnboxedMintToDouble() {
 | 
| -#if defined(TARGET_ARCH_IA32)
 | 
| -  return true;
 | 
| -#else
 | 
| -  // ARM does not have a short instruction sequence for converting int64 to
 | 
| -  // double.
 | 
| -  // TODO(johnmccutchan): Investigate possibility on MIPS once
 | 
| -  // mints are implemented there.
 | 
| -  return false;
 | 
| -#endif
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Optimize instance calls using ICData.
 | 
| -void FlowGraphOptimizer::ApplyICData() {
 | 
| -  VisitBlocks();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Optimize instance calls using cid.  This is called after optimizer
 | 
| -// converted instance calls to instructions. Any remaining
 | 
| -// instance calls are either megamorphic calls, cannot be optimized or
 | 
| -// have no runtime type feedback collected.
 | 
| -// Attempts to convert an instance call (IC call) using propagated class-ids,
 | 
| -// e.g., receiver class id, guarded-cid, or by guessing cid-s.
 | 
| -void FlowGraphOptimizer::ApplyClassIds() {
 | 
| -  ASSERT(current_iterator_ == NULL);
 | 
| -  for (intptr_t i = 0; i < block_order_.length(); ++i) {
 | 
| -    BlockEntryInstr* entry = block_order_[i];
 | 
| -    ForwardInstructionIterator it(entry);
 | 
| -    current_iterator_ = ⁢
 | 
| -    for (; !it.Done(); it.Advance()) {
 | 
| -      Instruction* instr = it.Current();
 | 
| -      if (instr->IsInstanceCall()) {
 | 
| -        InstanceCallInstr* call = instr->AsInstanceCall();
 | 
| -        if (call->HasICData()) {
 | 
| -          if (TryCreateICData(call)) {
 | 
| -            VisitInstanceCall(call);
 | 
| -          }
 | 
| -        }
 | 
| -      } else if (instr->IsPolymorphicInstanceCall()) {
 | 
| -        SpecializePolymorphicInstanceCall(instr->AsPolymorphicInstanceCall());
 | 
| -      } else if (instr->IsStrictCompare()) {
 | 
| -        VisitStrictCompare(instr->AsStrictCompare());
 | 
| -      } else if (instr->IsBranch()) {
 | 
| -        ComparisonInstr* compare = instr->AsBranch()->comparison();
 | 
| -        if (compare->IsStrictCompare()) {
 | 
| -          VisitStrictCompare(compare->AsStrictCompare());
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -    current_iterator_ = NULL;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// TODO(srdjan): Test/support other number types as well.
 | 
| -static bool IsNumberCid(intptr_t cid) {
 | 
| -  return (cid == kSmiCid) || (cid == kDoubleCid);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Attempt to build ICData for call using propagated class-ids.
 | 
| -bool FlowGraphOptimizer::TryCreateICData(InstanceCallInstr* call) {
 | 
| -  ASSERT(call->HasICData());
 | 
| -  if (call->ic_data()->NumberOfUsedChecks() > 0) {
 | 
| -    // This occurs when an instance call has too many checks, will be converted
 | 
| -    // to megamorphic call.
 | 
| -    return false;
 | 
| -  }
 | 
| -  if (FLAG_warn_on_javascript_compatibility) {
 | 
| -    // Do not make the instance call megamorphic if the callee needs to decode
 | 
| -    // the calling code sequence to lookup the ic data and verify if a warning
 | 
| -    // has already been issued or not.
 | 
| -    // TryCreateICData is only invoked if the ic_data target has not been called
 | 
| -    // yet, so no warning can possibly have been issued.
 | 
| -    ASSERT(!call->ic_data()->IssuedJSWarning());
 | 
| -    if (call->ic_data()->MayCheckForJSWarning()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -  GrowableArray<intptr_t> class_ids(call->ic_data()->NumArgsTested());
 | 
| -  ASSERT(call->ic_data()->NumArgsTested() <= call->ArgumentCount());
 | 
| -  for (intptr_t i = 0; i < call->ic_data()->NumArgsTested(); i++) {
 | 
| -    const intptr_t cid = call->PushArgumentAt(i)->value()->Type()->ToCid();
 | 
| -    class_ids.Add(cid);
 | 
| -  }
 | 
| -
 | 
| -  const Token::Kind op_kind = call->token_kind();
 | 
| -  if (Token::IsRelationalOperator(op_kind) ||
 | 
| -      Token::IsEqualityOperator(op_kind) ||
 | 
| -      Token::IsBinaryOperator(op_kind)) {
 | 
| -    // Guess cid: if one of the inputs is a number assume that the other
 | 
| -    // is a number of same type.
 | 
| -    const intptr_t cid_0 = class_ids[0];
 | 
| -    const intptr_t cid_1 = class_ids[1];
 | 
| -    if ((cid_0 == kDynamicCid) && (IsNumberCid(cid_1))) {
 | 
| -      class_ids[0] = cid_1;
 | 
| -    } else if (IsNumberCid(cid_0) && (cid_1 == kDynamicCid)) {
 | 
| -      class_ids[1] = cid_0;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  for (intptr_t i = 0; i < class_ids.length(); i++) {
 | 
| -    if (class_ids[i] == kDynamicCid) {
 | 
| -      // Not all cid-s known.
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  const Array& args_desc_array = Array::Handle(I,
 | 
| -      ArgumentsDescriptor::New(call->ArgumentCount(), call->argument_names()));
 | 
| -  ArgumentsDescriptor args_desc(args_desc_array);
 | 
| -  const Class& receiver_class = Class::Handle(I,
 | 
| -      isolate()->class_table()->At(class_ids[0]));
 | 
| -  const Function& function = Function::Handle(I,
 | 
| -      Resolver::ResolveDynamicForReceiverClass(
 | 
| -          receiver_class,
 | 
| -          call->function_name(),
 | 
| -          args_desc));
 | 
| -  if (function.IsNull()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  // Create new ICData, do not modify the one attached to the instruction
 | 
| -  // since it is attached to the assembly instruction itself.
 | 
| -  // TODO(srdjan): Prevent modification of ICData object that is
 | 
| -  // referenced in assembly code.
 | 
| -  ICData& ic_data = ICData::ZoneHandle(I, ICData::New(
 | 
| -      flow_graph_->parsed_function()->function(),
 | 
| -      call->function_name(),
 | 
| -      args_desc_array,
 | 
| -      call->deopt_id(),
 | 
| -      class_ids.length()));
 | 
| -  if (class_ids.length() > 1) {
 | 
| -    ic_data.AddCheck(class_ids, function);
 | 
| -  } else {
 | 
| -    ASSERT(class_ids.length() == 1);
 | 
| -    ic_data.AddReceiverCheck(class_ids[0], function);
 | 
| -  }
 | 
| -  call->set_ic_data(&ic_data);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -const ICData& FlowGraphOptimizer::TrySpecializeICData(const ICData& ic_data,
 | 
| -                                                      intptr_t cid) {
 | 
| -  ASSERT(ic_data.NumArgsTested() == 1);
 | 
| -
 | 
| -  if ((ic_data.NumberOfUsedChecks() == 1) && ic_data.HasReceiverClassId(cid)) {
 | 
| -    return ic_data;  // Nothing to do
 | 
| -  }
 | 
| -
 | 
| -  const Function& function =
 | 
| -      Function::Handle(I, ic_data.GetTargetForReceiverClassId(cid));
 | 
| -  // TODO(fschneider): Try looking up the function on the class if it is
 | 
| -  // not found in the ICData.
 | 
| -  if (!function.IsNull()) {
 | 
| -    const ICData& new_ic_data = ICData::ZoneHandle(I, ICData::New(
 | 
| -        Function::Handle(I, ic_data.owner()),
 | 
| -        String::Handle(I, ic_data.target_name()),
 | 
| -        Object::empty_array(),  // Dummy argument descriptor.
 | 
| -        ic_data.deopt_id(),
 | 
| -        ic_data.NumArgsTested()));
 | 
| -    new_ic_data.SetDeoptReasons(ic_data.DeoptReasons());
 | 
| -    new_ic_data.AddReceiverCheck(cid, function);
 | 
| -    return new_ic_data;
 | 
| -  }
 | 
| -
 | 
| -  return ic_data;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::SpecializePolymorphicInstanceCall(
 | 
| -    PolymorphicInstanceCallInstr* call) {
 | 
| -  if (!call->with_checks()) {
 | 
| -    return;  // Already specialized.
 | 
| -  }
 | 
| -
 | 
| -  const intptr_t receiver_cid =
 | 
| -      call->PushArgumentAt(0)->value()->Type()->ToCid();
 | 
| -  if (receiver_cid == kDynamicCid) {
 | 
| -    return;  // No information about receiver was infered.
 | 
| -  }
 | 
| -
 | 
| -  const ICData& ic_data = TrySpecializeICData(call->ic_data(), receiver_cid);
 | 
| -  if (ic_data.raw() == call->ic_data().raw()) {
 | 
| -    // No specialization.
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  const bool with_checks = false;
 | 
| -  PolymorphicInstanceCallInstr* specialized =
 | 
| -      new(I) PolymorphicInstanceCallInstr(call->instance_call(),
 | 
| -                                          ic_data,
 | 
| -                                          with_checks);
 | 
| -  call->ReplaceWith(specialized, current_iterator());
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static BinarySmiOpInstr* AsSmiShiftLeftInstruction(Definition* d) {
 | 
| -  BinarySmiOpInstr* instr = d->AsBinarySmiOp();
 | 
| -  if ((instr != NULL) && (instr->op_kind() == Token::kSHL)) {
 | 
| -    return instr;
 | 
| -  }
 | 
| -  return NULL;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool IsPositiveOrZeroSmiConst(Definition* d) {
 | 
| -  ConstantInstr* const_instr = d->AsConstant();
 | 
| -  if ((const_instr != NULL) && (const_instr->value().IsSmi())) {
 | 
| -    return Smi::Cast(const_instr->value()).Value() >= 0;
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::OptimizeLeftShiftBitAndSmiOp(
 | 
| -    Definition* bit_and_instr,
 | 
| -    Definition* left_instr,
 | 
| -    Definition* right_instr) {
 | 
| -  ASSERT(bit_and_instr != NULL);
 | 
| -  ASSERT((left_instr != NULL) && (right_instr != NULL));
 | 
| -
 | 
| -  // Check for pattern, smi_shift_left must be single-use.
 | 
| -  bool is_positive_or_zero = IsPositiveOrZeroSmiConst(left_instr);
 | 
| -  if (!is_positive_or_zero) {
 | 
| -    is_positive_or_zero = IsPositiveOrZeroSmiConst(right_instr);
 | 
| -  }
 | 
| -  if (!is_positive_or_zero) return;
 | 
| -
 | 
| -  BinarySmiOpInstr* smi_shift_left = NULL;
 | 
| -  if (bit_and_instr->InputAt(0)->IsSingleUse()) {
 | 
| -    smi_shift_left = AsSmiShiftLeftInstruction(left_instr);
 | 
| -  }
 | 
| -  if ((smi_shift_left == NULL) && (bit_and_instr->InputAt(1)->IsSingleUse())) {
 | 
| -    smi_shift_left = AsSmiShiftLeftInstruction(right_instr);
 | 
| -  }
 | 
| -  if (smi_shift_left == NULL) return;
 | 
| -
 | 
| -  // Pattern recognized.
 | 
| -  smi_shift_left->mark_truncating();
 | 
| -  ASSERT(bit_and_instr->IsBinarySmiOp() || bit_and_instr->IsBinaryMintOp());
 | 
| -  if (bit_and_instr->IsBinaryMintOp()) {
 | 
| -    // Replace Mint op with Smi op.
 | 
| -    BinarySmiOpInstr* smi_op = new(I) BinarySmiOpInstr(
 | 
| -        Token::kBIT_AND,
 | 
| -        new(I) Value(left_instr),
 | 
| -        new(I) Value(right_instr),
 | 
| -        Isolate::kNoDeoptId);  // BIT_AND cannot deoptimize.
 | 
| -    bit_and_instr->ReplaceWith(smi_op, current_iterator());
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -
 | 
| -// Used by TryMergeDivMod.
 | 
| -// Inserts a load-indexed instruction between a TRUNCDIV or MOD instruction,
 | 
| -// and the using instruction. This is an intermediate step before merging.
 | 
| -void FlowGraphOptimizer::AppendLoadIndexedForMerged(Definition* instr,
 | 
| -                                                    intptr_t ix,
 | 
| -                                                    intptr_t cid) {
 | 
| -  const intptr_t index_scale = Instance::ElementSizeFor(cid);
 | 
| -  ConstantInstr* index_instr =
 | 
| -      flow_graph()->GetConstant(Smi::Handle(I, Smi::New(ix)));
 | 
| -  LoadIndexedInstr* load =
 | 
| -      new(I) LoadIndexedInstr(new(I) Value(instr),
 | 
| -                                      new(I) Value(index_instr),
 | 
| -                                      index_scale,
 | 
| -                                      cid,
 | 
| -                                      Isolate::kNoDeoptId,
 | 
| -                                      instr->token_pos());
 | 
| -  instr->ReplaceUsesWith(load);
 | 
| -  flow_graph()->InsertAfter(instr, load, NULL, FlowGraph::kValue);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::AppendExtractNthOutputForMerged(Definition* instr,
 | 
| -                                                         intptr_t index,
 | 
| -                                                         Representation rep,
 | 
| -                                                         intptr_t cid) {
 | 
| -  ExtractNthOutputInstr* extract =
 | 
| -      new(I) ExtractNthOutputInstr(new(I) Value(instr), index, rep, cid);
 | 
| -  instr->ReplaceUsesWith(extract);
 | 
| -  flow_graph()->InsertAfter(instr, extract, NULL, FlowGraph::kValue);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Dart:
 | 
| -//  var x = d % 10;
 | 
| -//  var y = d ~/ 10;
 | 
| -//  var z = x + y;
 | 
| -//
 | 
| -// IL:
 | 
| -//  v4 <- %(v2, v3)
 | 
| -//  v5 <- ~/(v2, v3)
 | 
| -//  v6 <- +(v4, v5)
 | 
| -//
 | 
| -// IL optimized:
 | 
| -//  v4 <- DIVMOD(v2, v3);
 | 
| -//  v5 <- LoadIndexed(v4, 0); // ~/ result
 | 
| -//  v6 <- LoadIndexed(v4, 1); // % result
 | 
| -//  v7 <- +(v5, v6)
 | 
| -// Because of the environment it is important that merged instruction replaces
 | 
| -// first original instruction encountered.
 | 
| -void FlowGraphOptimizer::TryMergeTruncDivMod(
 | 
| -    GrowableArray<BinarySmiOpInstr*>* merge_candidates) {
 | 
| -  if (merge_candidates->length() < 2) {
 | 
| -    // Need at least a TRUNCDIV and a MOD.
 | 
| -    return;
 | 
| -  }
 | 
| -  for (intptr_t i = 0; i < merge_candidates->length(); i++) {
 | 
| -    BinarySmiOpInstr* curr_instr = (*merge_candidates)[i];
 | 
| -    if (curr_instr == NULL) {
 | 
| -      // Instruction was merged already.
 | 
| -      continue;
 | 
| -    }
 | 
| -    ASSERT((curr_instr->op_kind() == Token::kTRUNCDIV) ||
 | 
| -           (curr_instr->op_kind() == Token::kMOD));
 | 
| -    // Check if there is kMOD/kTRUNDIV binop with same inputs.
 | 
| -    const intptr_t other_kind = (curr_instr->op_kind() == Token::kTRUNCDIV) ?
 | 
| -        Token::kMOD : Token::kTRUNCDIV;
 | 
| -    Definition* left_def = curr_instr->left()->definition();
 | 
| -    Definition* right_def = curr_instr->right()->definition();
 | 
| -    for (intptr_t k = i + 1; k < merge_candidates->length(); k++) {
 | 
| -      BinarySmiOpInstr* other_binop = (*merge_candidates)[k];
 | 
| -      // 'other_binop' can be NULL if it was already merged.
 | 
| -      if ((other_binop != NULL) &&
 | 
| -          (other_binop->op_kind() == other_kind) &&
 | 
| -          (other_binop->left()->definition() == left_def) &&
 | 
| -          (other_binop->right()->definition() == right_def)) {
 | 
| -        (*merge_candidates)[k] = NULL;  // Clear it.
 | 
| -        ASSERT(curr_instr->HasUses());
 | 
| -        AppendExtractNthOutputForMerged(
 | 
| -            curr_instr,
 | 
| -            MergedMathInstr::OutputIndexOf(curr_instr->op_kind()),
 | 
| -            kTagged, kSmiCid);
 | 
| -        ASSERT(other_binop->HasUses());
 | 
| -        AppendExtractNthOutputForMerged(
 | 
| -            other_binop,
 | 
| -            MergedMathInstr::OutputIndexOf(other_binop->op_kind()),
 | 
| -            kTagged, kSmiCid);
 | 
| -
 | 
| -        ZoneGrowableArray<Value*>* args = new(I) ZoneGrowableArray<Value*>(2);
 | 
| -        args->Add(new(I) Value(curr_instr->left()->definition()));
 | 
| -        args->Add(new(I) Value(curr_instr->right()->definition()));
 | 
| -
 | 
| -        // Replace with TruncDivMod.
 | 
| -        MergedMathInstr* div_mod = new(I) MergedMathInstr(
 | 
| -            args,
 | 
| -            curr_instr->deopt_id(),
 | 
| -            MergedMathInstr::kTruncDivMod);
 | 
| -        curr_instr->ReplaceWith(div_mod, current_iterator());
 | 
| -        other_binop->ReplaceUsesWith(div_mod);
 | 
| -        other_binop->RemoveFromGraph();
 | 
| -        // Only one merge possible. Because canonicalization happens later,
 | 
| -        // more candidates are possible.
 | 
| -        // TODO(srdjan): Allow merging of trunc-div/mod into truncDivMod.
 | 
| -        break;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Tries to merge MathUnary operations, in this case sinus and cosinus.
 | 
| -void FlowGraphOptimizer::TryMergeMathUnary(
 | 
| -    GrowableArray<MathUnaryInstr*>* merge_candidates) {
 | 
| -  if (!FlowGraphCompiler::SupportsSinCos() || !CanUnboxDouble()) {
 | 
| -    return;
 | 
| -  }
 | 
| -  if (merge_candidates->length() < 2) {
 | 
| -    // Need at least a SIN and a COS.
 | 
| -    return;
 | 
| -  }
 | 
| -  for (intptr_t i = 0; i < merge_candidates->length(); i++) {
 | 
| -    MathUnaryInstr* curr_instr = (*merge_candidates)[i];
 | 
| -    if (curr_instr == NULL) {
 | 
| -      // Instruction was merged already.
 | 
| -      continue;
 | 
| -    }
 | 
| -    const intptr_t kind = curr_instr->kind();
 | 
| -    ASSERT((kind == MathUnaryInstr::kSin) ||
 | 
| -           (kind == MathUnaryInstr::kCos));
 | 
| -    // Check if there is sin/cos binop with same inputs.
 | 
| -    const intptr_t other_kind = (kind == MethodRecognizer::kMathSin) ?
 | 
| -        MethodRecognizer::kMathCos : MethodRecognizer::kMathSin;
 | 
| -    Definition* def = curr_instr->value()->definition();
 | 
| -    for (intptr_t k = i + 1; k < merge_candidates->length(); k++) {
 | 
| -      MathUnaryInstr* other_op = (*merge_candidates)[k];
 | 
| -      // 'other_op' can be NULL if it was already merged.
 | 
| -      if ((other_op != NULL) && (other_op->kind() == other_kind) &&
 | 
| -          (other_op->value()->definition() == def)) {
 | 
| -        (*merge_candidates)[k] = NULL;  // Clear it.
 | 
| -        ASSERT(curr_instr->HasUses());
 | 
| -        AppendExtractNthOutputForMerged(curr_instr,
 | 
| -                                        MergedMathInstr::OutputIndexOf(kind),
 | 
| -                                        kUnboxedDouble, kDoubleCid);
 | 
| -        ASSERT(other_op->HasUses());
 | 
| -        AppendExtractNthOutputForMerged(
 | 
| -            other_op,
 | 
| -            MergedMathInstr::OutputIndexOf(other_kind),
 | 
| -            kUnboxedDouble, kDoubleCid);
 | 
| -        ZoneGrowableArray<Value*>* args = new(I) ZoneGrowableArray<Value*>(1);
 | 
| -        args->Add(new(I) Value(curr_instr->value()->definition()));
 | 
| -        // Replace with SinCos.
 | 
| -        MergedMathInstr* sin_cos =
 | 
| -            new(I) MergedMathInstr(args,
 | 
| -                                   curr_instr->DeoptimizationTarget(),
 | 
| -                                   MergedMathInstr::kSinCos);
 | 
| -        curr_instr->ReplaceWith(sin_cos, current_iterator());
 | 
| -        other_op->ReplaceUsesWith(sin_cos);
 | 
| -        other_op->RemoveFromGraph();
 | 
| -        // Only one merge possible. Because canonicalization happens later,
 | 
| -        // more candidates are possible.
 | 
| -        // TODO(srdjan): Allow merging of sin/cos into sincos.
 | 
| -        break;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Optimize (a << b) & c pattern: if c is a positive Smi or zero, then the
 | 
| -// shift can be a truncating Smi shift-left and result is always Smi.
 | 
| -// Merging occurs only per basic-block.
 | 
| -void FlowGraphOptimizer::TryOptimizePatterns() {
 | 
| -  if (!FLAG_truncating_left_shift) return;
 | 
| -  ASSERT(current_iterator_ == NULL);
 | 
| -  GrowableArray<BinarySmiOpInstr*> div_mod_merge;
 | 
| -  GrowableArray<MathUnaryInstr*> sin_cos_merge;
 | 
| -  for (intptr_t i = 0; i < block_order_.length(); ++i) {
 | 
| -    // Merging only per basic-block.
 | 
| -    div_mod_merge.Clear();
 | 
| -    sin_cos_merge.Clear();
 | 
| -    BlockEntryInstr* entry = block_order_[i];
 | 
| -    ForwardInstructionIterator it(entry);
 | 
| -    current_iterator_ = ⁢
 | 
| -    for (; !it.Done(); it.Advance()) {
 | 
| -      if (it.Current()->IsBinarySmiOp()) {
 | 
| -        BinarySmiOpInstr* binop = it.Current()->AsBinarySmiOp();
 | 
| -        if (binop->op_kind() == Token::kBIT_AND) {
 | 
| -          OptimizeLeftShiftBitAndSmiOp(binop,
 | 
| -                                       binop->left()->definition(),
 | 
| -                                       binop->right()->definition());
 | 
| -        } else if ((binop->op_kind() == Token::kTRUNCDIV) ||
 | 
| -                   (binop->op_kind() == Token::kMOD)) {
 | 
| -          if (binop->HasUses()) {
 | 
| -            div_mod_merge.Add(binop);
 | 
| -          }
 | 
| -        }
 | 
| -      } else if (it.Current()->IsBinaryMintOp()) {
 | 
| -        BinaryMintOpInstr* mintop = it.Current()->AsBinaryMintOp();
 | 
| -        if (mintop->op_kind() == Token::kBIT_AND) {
 | 
| -          OptimizeLeftShiftBitAndSmiOp(mintop,
 | 
| -                                       mintop->left()->definition(),
 | 
| -                                       mintop->right()->definition());
 | 
| -        }
 | 
| -      } else if (it.Current()->IsMathUnary()) {
 | 
| -        MathUnaryInstr* math_unary = it.Current()->AsMathUnary();
 | 
| -        if ((math_unary->kind() == MathUnaryInstr::kSin) ||
 | 
| -            (math_unary->kind() == MathUnaryInstr::kCos)) {
 | 
| -          if (math_unary->HasUses()) {
 | 
| -            sin_cos_merge.Add(math_unary);
 | 
| -          }
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -    TryMergeTruncDivMod(&div_mod_merge);
 | 
| -    TryMergeMathUnary(&sin_cos_merge);
 | 
| -    current_iterator_ = NULL;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static void EnsureSSATempIndex(FlowGraph* graph,
 | 
| -                               Definition* defn,
 | 
| -                               Definition* replacement) {
 | 
| -  if ((replacement->ssa_temp_index() == -1) &&
 | 
| -      (defn->ssa_temp_index() != -1)) {
 | 
| -    replacement->set_ssa_temp_index(graph->alloc_ssa_temp_index());
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static void ReplaceCurrentInstruction(ForwardInstructionIterator* iterator,
 | 
| -                                      Instruction* current,
 | 
| -                                      Instruction* replacement,
 | 
| -                                      FlowGraph* graph) {
 | 
| -  Definition* current_defn = current->AsDefinition();
 | 
| -  if ((replacement != NULL) && (current_defn != NULL)) {
 | 
| -    Definition* replacement_defn = replacement->AsDefinition();
 | 
| -    ASSERT(replacement_defn != NULL);
 | 
| -    current_defn->ReplaceUsesWith(replacement_defn);
 | 
| -    EnsureSSATempIndex(graph, current_defn, replacement_defn);
 | 
| -
 | 
| -    if (FLAG_trace_optimization) {
 | 
| -      OS::Print("Replacing v%" Pd " with v%" Pd "\n",
 | 
| -                current_defn->ssa_temp_index(),
 | 
| -                replacement_defn->ssa_temp_index());
 | 
| -    }
 | 
| -  } else if (FLAG_trace_optimization) {
 | 
| -    if (current_defn == NULL) {
 | 
| -      OS::Print("Removing %s\n", current->DebugName());
 | 
| -    } else {
 | 
| -      ASSERT(!current_defn->HasUses());
 | 
| -      OS::Print("Removing v%" Pd ".\n", current_defn->ssa_temp_index());
 | 
| -    }
 | 
| -  }
 | 
| -  iterator->RemoveCurrentFromGraph();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::Canonicalize() {
 | 
| -  bool changed = false;
 | 
| -  for (intptr_t i = 0; i < block_order_.length(); ++i) {
 | 
| -    BlockEntryInstr* entry = block_order_[i];
 | 
| -    for (ForwardInstructionIterator it(entry); !it.Done(); it.Advance()) {
 | 
| -      Instruction* current = it.Current();
 | 
| -      if (current->HasUnmatchedInputRepresentations()) {
 | 
| -        // Can't canonicalize this instruction until all conversions for its
 | 
| -        // inputs are inserted.
 | 
| -        continue;
 | 
| -      }
 | 
| -
 | 
| -      Instruction* replacement = current->Canonicalize(flow_graph());
 | 
| -
 | 
| -      if (replacement != current) {
 | 
| -        // For non-definitions Canonicalize should return either NULL or
 | 
| -        // this.
 | 
| -        ASSERT((replacement == NULL) || current->IsDefinition());
 | 
| -        ReplaceCurrentInstruction(&it, current, replacement, flow_graph_);
 | 
| -        changed = true;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -  return changed;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool IsUnboxedInteger(Representation rep) {
 | 
| -  return (rep == kUnboxedInt32) ||
 | 
| -         (rep == kUnboxedUint32) ||
 | 
| -         (rep == kUnboxedMint);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::InsertConversion(Representation from,
 | 
| -                                          Representation to,
 | 
| -                                          Value* use,
 | 
| -                                          bool is_environment_use) {
 | 
| -  Instruction* insert_before;
 | 
| -  Instruction* deopt_target;
 | 
| -  PhiInstr* phi = use->instruction()->AsPhi();
 | 
| -  if (phi != NULL) {
 | 
| -    ASSERT(phi->is_alive());
 | 
| -    // For phis conversions have to be inserted in the predecessor.
 | 
| -    insert_before =
 | 
| -        phi->block()->PredecessorAt(use->use_index())->last_instruction();
 | 
| -    deopt_target = NULL;
 | 
| -  } else {
 | 
| -    deopt_target = insert_before = use->instruction();
 | 
| -  }
 | 
| -
 | 
| -  Definition* converted = NULL;
 | 
| -  if (IsUnboxedInteger(from) && IsUnboxedInteger(to)) {
 | 
| -    const intptr_t deopt_id = (to == kUnboxedInt32) && (deopt_target != NULL) ?
 | 
| -      deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId;
 | 
| -    converted = new(I) UnboxedIntConverterInstr(from,
 | 
| -                                                to,
 | 
| -                                                use->CopyWithType(),
 | 
| -                                                deopt_id);
 | 
| -  } else if ((from == kUnboxedInt32) && (to == kUnboxedDouble)) {
 | 
| -    converted = new Int32ToDoubleInstr(use->CopyWithType());
 | 
| -  } else if ((from == kUnboxedMint) &&
 | 
| -             (to == kUnboxedDouble) &&
 | 
| -             CanConvertUnboxedMintToDouble()) {
 | 
| -    const intptr_t deopt_id = (deopt_target != NULL) ?
 | 
| -      deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId;
 | 
| -    ASSERT(CanUnboxDouble());
 | 
| -    converted = new MintToDoubleInstr(use->CopyWithType(), deopt_id);
 | 
| -  } else if ((from == kTagged) && Boxing::Supports(to)) {
 | 
| -    const intptr_t deopt_id = (deopt_target != NULL) ?
 | 
| -      deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId;
 | 
| -    converted = UnboxInstr::Create(to, use->CopyWithType(), deopt_id);
 | 
| -  } else if ((to == kTagged) && Boxing::Supports(from)) {
 | 
| -    converted = BoxInstr::Create(from, use->CopyWithType());
 | 
| -  } else {
 | 
| -    // We have failed to find a suitable conversion instruction.
 | 
| -    // Insert two "dummy" conversion instructions with the correct
 | 
| -    // "from" and "to" representation. The inserted instructions will
 | 
| -    // trigger a deoptimization if executed. See #12417 for a discussion.
 | 
| -    const intptr_t deopt_id = (deopt_target != NULL) ?
 | 
| -      deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId;
 | 
| -    ASSERT(Boxing::Supports(from));
 | 
| -    ASSERT(Boxing::Supports(to));
 | 
| -    Definition* boxed = BoxInstr::Create(from, use->CopyWithType());
 | 
| -    use->BindTo(boxed);
 | 
| -    InsertBefore(insert_before, boxed, NULL, FlowGraph::kValue);
 | 
| -    converted = UnboxInstr::Create(to, new(I) Value(boxed), deopt_id);
 | 
| -  }
 | 
| -  ASSERT(converted != NULL);
 | 
| -  InsertBefore(insert_before, converted, use->instruction()->env(),
 | 
| -               FlowGraph::kValue);
 | 
| -  if (is_environment_use) {
 | 
| -    use->BindToEnvironment(converted);
 | 
| -  } else {
 | 
| -    use->BindTo(converted);
 | 
| -  }
 | 
| -
 | 
| -  if ((to == kUnboxedInt32) && (phi != NULL)) {
 | 
| -    // Int32 phis are unboxed optimistically. Ensure that unboxing
 | 
| -    // has deoptimization target attached from the goto instruction.
 | 
| -    flow_graph_->CopyDeoptTarget(converted, insert_before);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::ConvertUse(Value* use, Representation from_rep) {
 | 
| -  const Representation to_rep =
 | 
| -      use->instruction()->RequiredInputRepresentation(use->use_index());
 | 
| -  if (from_rep == to_rep || to_rep == kNoRepresentation) {
 | 
| -    return;
 | 
| -  }
 | 
| -  InsertConversion(from_rep, to_rep, use, /*is_environment_use=*/ false);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::ConvertEnvironmentUse(Value* use,
 | 
| -                                               Representation from_rep) {
 | 
| -  const Representation to_rep = kTagged;
 | 
| -  if (from_rep == to_rep || to_rep == kNoRepresentation) {
 | 
| -    return;
 | 
| -  }
 | 
| -  InsertConversion(from_rep, to_rep, use, /*is_environment_use=*/ true);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::InsertConversionsFor(Definition* def) {
 | 
| -  const Representation from_rep = def->representation();
 | 
| -
 | 
| -  for (Value::Iterator it(def->input_use_list());
 | 
| -       !it.Done();
 | 
| -       it.Advance()) {
 | 
| -    ConvertUse(it.Current(), from_rep);
 | 
| -  }
 | 
| -
 | 
| -  if (flow_graph()->graph_entry()->SuccessorCount() > 1) {
 | 
| -    for (Value::Iterator it(def->env_use_list());
 | 
| -         !it.Done();
 | 
| -         it.Advance()) {
 | 
| -      Value* use = it.Current();
 | 
| -      if (use->instruction()->MayThrow() &&
 | 
| -          use->instruction()->GetBlock()->InsideTryBlock()) {
 | 
| -        // Environment uses at calls inside try-blocks must be converted to
 | 
| -        // tagged representation.
 | 
| -        ConvertEnvironmentUse(it.Current(), from_rep);
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static void UnboxPhi(PhiInstr* phi) {
 | 
| -  Representation unboxed = phi->representation();
 | 
| -
 | 
| -  switch (phi->Type()->ToCid()) {
 | 
| -    case kDoubleCid:
 | 
| -      if (CanUnboxDouble()) {
 | 
| -        unboxed = kUnboxedDouble;
 | 
| -      }
 | 
| -      break;
 | 
| -    case kFloat32x4Cid:
 | 
| -      if (ShouldInlineSimd()) {
 | 
| -        unboxed = kUnboxedFloat32x4;
 | 
| -      }
 | 
| -      break;
 | 
| -    case kInt32x4Cid:
 | 
| -      if (ShouldInlineSimd()) {
 | 
| -        unboxed = kUnboxedInt32x4;
 | 
| -      }
 | 
| -      break;
 | 
| -    case kFloat64x2Cid:
 | 
| -      if (ShouldInlineSimd()) {
 | 
| -        unboxed = kUnboxedFloat64x2;
 | 
| -      }
 | 
| -      break;
 | 
| -  }
 | 
| -
 | 
| -  if ((kSmiBits < 32) &&
 | 
| -      (unboxed == kTagged) &&
 | 
| -      phi->Type()->IsInt() &&
 | 
| -      RangeUtils::Fits(phi->range(), RangeBoundary::kRangeBoundaryInt32)) {
 | 
| -    // On 32-bit platforms conservatively unbox phis that:
 | 
| -    //   - are proven to be of type Int;
 | 
| -    //   - fit into 32bits range;
 | 
| -    //   - have either constants or Box() operations as inputs;
 | 
| -    //   - have at least one Box() operation as an input;
 | 
| -    //   - are used in at least 1 Unbox() operation.
 | 
| -    bool should_unbox = false;
 | 
| -    for (intptr_t i = 0; i < phi->InputCount(); i++) {
 | 
| -      Definition* input = phi->InputAt(i)->definition();
 | 
| -      if (input->IsBox() &&
 | 
| -          RangeUtils::Fits(input->range(),
 | 
| -                           RangeBoundary::kRangeBoundaryInt32)) {
 | 
| -        should_unbox = true;
 | 
| -      } else if (!input->IsConstant()) {
 | 
| -        should_unbox = false;
 | 
| -        break;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    if (should_unbox) {
 | 
| -      // We checked inputs. Check if phi is used in at least one unbox
 | 
| -      // operation.
 | 
| -      bool has_unboxed_use = false;
 | 
| -      for (Value* use = phi->input_use_list();
 | 
| -           use != NULL;
 | 
| -           use = use->next_use()) {
 | 
| -        Instruction* instr = use->instruction();
 | 
| -        if (instr->IsUnbox()) {
 | 
| -          has_unboxed_use = true;
 | 
| -          break;
 | 
| -        } else if (IsUnboxedInteger(
 | 
| -            instr->RequiredInputRepresentation(use->use_index()))) {
 | 
| -          has_unboxed_use = true;
 | 
| -          break;
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      if (!has_unboxed_use) {
 | 
| -        should_unbox = false;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    if (should_unbox) {
 | 
| -      unboxed = kUnboxedInt32;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  phi->set_representation(unboxed);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::SelectRepresentations() {
 | 
| -  // Conservatively unbox all phis that were proven to be of Double,
 | 
| -  // Float32x4, or Int32x4 type.
 | 
| -  for (intptr_t i = 0; i < block_order_.length(); ++i) {
 | 
| -    JoinEntryInstr* join_entry = block_order_[i]->AsJoinEntry();
 | 
| -    if (join_entry != NULL) {
 | 
| -      for (PhiIterator it(join_entry); !it.Done(); it.Advance()) {
 | 
| -        PhiInstr* phi = it.Current();
 | 
| -        UnboxPhi(phi);
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Process all instructions and insert conversions where needed.
 | 
| -  GraphEntryInstr* graph_entry = block_order_[0]->AsGraphEntry();
 | 
| -
 | 
| -  // Visit incoming parameters and constants.
 | 
| -  for (intptr_t i = 0; i < graph_entry->initial_definitions()->length(); i++) {
 | 
| -    InsertConversionsFor((*graph_entry->initial_definitions())[i]);
 | 
| -  }
 | 
| -
 | 
| -  for (intptr_t i = 0; i < block_order_.length(); ++i) {
 | 
| -    BlockEntryInstr* entry = block_order_[i];
 | 
| -    JoinEntryInstr* join_entry = entry->AsJoinEntry();
 | 
| -    if (join_entry != NULL) {
 | 
| -      for (PhiIterator it(join_entry); !it.Done(); it.Advance()) {
 | 
| -        PhiInstr* phi = it.Current();
 | 
| -        ASSERT(phi != NULL);
 | 
| -        ASSERT(phi->is_alive());
 | 
| -        InsertConversionsFor(phi);
 | 
| -      }
 | 
| -    }
 | 
| -    CatchBlockEntryInstr* catch_entry = entry->AsCatchBlockEntry();
 | 
| -    if (catch_entry != NULL) {
 | 
| -      for (intptr_t i = 0;
 | 
| -           i < catch_entry->initial_definitions()->length();
 | 
| -           i++) {
 | 
| -        InsertConversionsFor((*catch_entry->initial_definitions())[i]);
 | 
| -      }
 | 
| -    }
 | 
| -    for (ForwardInstructionIterator it(entry); !it.Done(); it.Advance()) {
 | 
| -      Definition* def = it.Current()->AsDefinition();
 | 
| -      if (def != NULL) {
 | 
| -        InsertConversionsFor(def);
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool ClassIdIsOneOf(intptr_t class_id,
 | 
| -                           const GrowableArray<intptr_t>& class_ids) {
 | 
| -  for (intptr_t i = 0; i < class_ids.length(); i++) {
 | 
| -    ASSERT(class_ids[i] != kIllegalCid);
 | 
| -    if (class_ids[i] == class_id) {
 | 
| -      return true;
 | 
| -    }
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Returns true if ICData tests two arguments and all ICData cids are in the
 | 
| -// required sets 'receiver_class_ids' or 'argument_class_ids', respectively.
 | 
| -static bool ICDataHasOnlyReceiverArgumentClassIds(
 | 
| -    const ICData& ic_data,
 | 
| -    const GrowableArray<intptr_t>& receiver_class_ids,
 | 
| -    const GrowableArray<intptr_t>& argument_class_ids) {
 | 
| -  if (ic_data.NumArgsTested() != 2) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  Function& target = Function::Handle();
 | 
| -  const intptr_t len = ic_data.NumberOfChecks();
 | 
| -  GrowableArray<intptr_t> class_ids;
 | 
| -  for (intptr_t i = 0; i < len; i++) {
 | 
| -    if (ic_data.IsUsedAt(i)) {
 | 
| -      ic_data.GetCheckAt(i, &class_ids, &target);
 | 
| -      ASSERT(class_ids.length() == 2);
 | 
| -      if (!ClassIdIsOneOf(class_ids[0], receiver_class_ids) ||
 | 
| -          !ClassIdIsOneOf(class_ids[1], argument_class_ids)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool ICDataHasReceiverArgumentClassIds(const ICData& ic_data,
 | 
| -                                              intptr_t receiver_class_id,
 | 
| -                                              intptr_t argument_class_id) {
 | 
| -  if (ic_data.NumArgsTested() != 2) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  Function& target = Function::Handle();
 | 
| -  const intptr_t len = ic_data.NumberOfChecks();
 | 
| -  for (intptr_t i = 0; i < len; i++) {
 | 
| -    if (ic_data.IsUsedAt(i)) {
 | 
| -      GrowableArray<intptr_t> class_ids;
 | 
| -      ic_data.GetCheckAt(i, &class_ids, &target);
 | 
| -      ASSERT(class_ids.length() == 2);
 | 
| -      if ((class_ids[0] == receiver_class_id) &&
 | 
| -          (class_ids[1] == argument_class_id)) {
 | 
| -        return true;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool HasOnlyOneSmi(const ICData& ic_data) {
 | 
| -  return (ic_data.NumberOfUsedChecks() == 1)
 | 
| -      && ic_data.HasReceiverClassId(kSmiCid);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool HasOnlySmiOrMint(const ICData& ic_data) {
 | 
| -  if (ic_data.NumberOfUsedChecks() == 1) {
 | 
| -    return ic_data.HasReceiverClassId(kSmiCid)
 | 
| -        || ic_data.HasReceiverClassId(kMintCid);
 | 
| -  }
 | 
| -  return (ic_data.NumberOfUsedChecks() == 2)
 | 
| -      && ic_data.HasReceiverClassId(kSmiCid)
 | 
| -      && ic_data.HasReceiverClassId(kMintCid);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool HasOnlyTwoOf(const ICData& ic_data, intptr_t cid) {
 | 
| -  if (ic_data.NumberOfUsedChecks() != 1) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  GrowableArray<intptr_t> first;
 | 
| -  GrowableArray<intptr_t> second;
 | 
| -  ic_data.GetUsedCidsForTwoArgs(&first, &second);
 | 
| -  return (first[0] == cid) && (second[0] == cid);
 | 
| -}
 | 
| -
 | 
| -// Returns false if the ICData contains anything other than the 4 combinations
 | 
| -// of Mint and Smi for the receiver and argument classes.
 | 
| -static bool HasTwoMintOrSmi(const ICData& ic_data) {
 | 
| -  GrowableArray<intptr_t> first;
 | 
| -  GrowableArray<intptr_t> second;
 | 
| -  ic_data.GetUsedCidsForTwoArgs(&first, &second);
 | 
| -  for (intptr_t i = 0; i < first.length(); i++) {
 | 
| -    if ((first[i] != kSmiCid) && (first[i] != kMintCid)) {
 | 
| -      return false;
 | 
| -    }
 | 
| -    if ((second[i] != kSmiCid) && (second[i] != kMintCid)) {
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Returns false if the ICData contains anything other than the 4 combinations
 | 
| -// of Double and Smi for the receiver and argument classes.
 | 
| -static bool HasTwoDoubleOrSmi(const ICData& ic_data) {
 | 
| -  GrowableArray<intptr_t> class_ids(2);
 | 
| -  class_ids.Add(kSmiCid);
 | 
| -  class_ids.Add(kDoubleCid);
 | 
| -  return ICDataHasOnlyReceiverArgumentClassIds(ic_data, class_ids, class_ids);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool HasOnlyOneDouble(const ICData& ic_data) {
 | 
| -  return (ic_data.NumberOfUsedChecks() == 1)
 | 
| -      && ic_data.HasReceiverClassId(kDoubleCid);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool ShouldSpecializeForDouble(const ICData& ic_data) {
 | 
| -  // Don't specialize for double if we can't unbox them.
 | 
| -  if (!CanUnboxDouble()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  // Unboxed double operation can't handle case of two smis.
 | 
| -  if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) {
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  // Check that it have seen only smis and doubles.
 | 
| -  return HasTwoDoubleOrSmi(ic_data);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::ReplaceCall(Definition* call,
 | 
| -                                     Definition* replacement) {
 | 
| -  // Remove the original push arguments.
 | 
| -  for (intptr_t i = 0; i < call->ArgumentCount(); ++i) {
 | 
| -    PushArgumentInstr* push = call->PushArgumentAt(i);
 | 
| -    push->ReplaceUsesWith(push->value()->definition());
 | 
| -    push->RemoveFromGraph();
 | 
| -  }
 | 
| -  call->ReplaceWith(replacement, current_iterator());
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::AddCheckSmi(Definition* to_check,
 | 
| -                                     intptr_t deopt_id,
 | 
| -                                     Environment* deopt_environment,
 | 
| -                                     Instruction* insert_before) {
 | 
| -  if (to_check->Type()->ToCid() != kSmiCid) {
 | 
| -    InsertBefore(insert_before,
 | 
| -                 new(I) CheckSmiInstr(new(I) Value(to_check),
 | 
| -                                      deopt_id,
 | 
| -                                      insert_before->token_pos()),
 | 
| -                 deopt_environment,
 | 
| -                 FlowGraph::kEffect);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -Instruction* FlowGraphOptimizer::GetCheckClass(Definition* to_check,
 | 
| -                                               const ICData& unary_checks,
 | 
| -                                               intptr_t deopt_id,
 | 
| -                                               intptr_t token_pos) {
 | 
| -  if ((unary_checks.NumberOfUsedChecks() == 1) &&
 | 
| -      unary_checks.HasReceiverClassId(kSmiCid)) {
 | 
| -    return new(I) CheckSmiInstr(new(I) Value(to_check),
 | 
| -                                deopt_id,
 | 
| -                                token_pos);
 | 
| -  }
 | 
| -  return new(I) CheckClassInstr(
 | 
| -      new(I) Value(to_check), deopt_id, unary_checks, token_pos);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::AddCheckClass(Definition* to_check,
 | 
| -                                       const ICData& unary_checks,
 | 
| -                                       intptr_t deopt_id,
 | 
| -                                       Environment* deopt_environment,
 | 
| -                                       Instruction* insert_before) {
 | 
| -  // Type propagation has not run yet, we cannot eliminate the check.
 | 
| -  Instruction* check = GetCheckClass(
 | 
| -      to_check, unary_checks, deopt_id, insert_before->token_pos());
 | 
| -  InsertBefore(insert_before, check, deopt_environment, FlowGraph::kEffect);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::AddReceiverCheck(InstanceCallInstr* call) {
 | 
| -  AddCheckClass(call->ArgumentAt(0),
 | 
| -                ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks()),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool ArgIsAlways(intptr_t cid,
 | 
| -                        const ICData& ic_data,
 | 
| -                        intptr_t arg_number) {
 | 
| -  ASSERT(ic_data.NumArgsTested() > arg_number);
 | 
| -  if (ic_data.NumberOfUsedChecks() == 0) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  const intptr_t num_checks = ic_data.NumberOfChecks();
 | 
| -  for (intptr_t i = 0; i < num_checks; i++) {
 | 
| -    if (ic_data.IsUsedAt(i) && ic_data.GetClassIdAt(i, arg_number) != cid) {
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool CanUnboxInt32() {
 | 
| -  // Int32/Uint32 can be unboxed if it fits into a smi or the platform
 | 
| -  // supports unboxed mints.
 | 
| -  return (kSmiBits >= 32) || FlowGraphCompiler::SupportsUnboxedMints();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static intptr_t MethodKindToCid(MethodRecognizer::Kind kind) {
 | 
| -  switch (kind) {
 | 
| -    case MethodRecognizer::kImmutableArrayGetIndexed:
 | 
| -      return kImmutableArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kObjectArrayGetIndexed:
 | 
| -    case MethodRecognizer::kObjectArraySetIndexed:
 | 
| -      return kArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kGrowableArrayGetIndexed:
 | 
| -    case MethodRecognizer::kGrowableArraySetIndexed:
 | 
| -      return kGrowableObjectArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kFloat32ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kFloat32ArraySetIndexed:
 | 
| -      return kTypedDataFloat32ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kFloat64ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kFloat64ArraySetIndexed:
 | 
| -      return kTypedDataFloat64ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kInt8ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kInt8ArraySetIndexed:
 | 
| -      return kTypedDataInt8ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kUint8ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint8ArraySetIndexed:
 | 
| -      return kTypedDataUint8ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kUint8ClampedArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint8ClampedArraySetIndexed:
 | 
| -      return kTypedDataUint8ClampedArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kExternalUint8ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kExternalUint8ArraySetIndexed:
 | 
| -      return kExternalTypedDataUint8ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kExternalUint8ClampedArrayGetIndexed:
 | 
| -    case MethodRecognizer::kExternalUint8ClampedArraySetIndexed:
 | 
| -      return kExternalTypedDataUint8ClampedArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kInt16ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kInt16ArraySetIndexed:
 | 
| -      return kTypedDataInt16ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kUint16ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint16ArraySetIndexed:
 | 
| -      return kTypedDataUint16ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kInt32ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kInt32ArraySetIndexed:
 | 
| -      return kTypedDataInt32ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kUint32ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint32ArraySetIndexed:
 | 
| -      return kTypedDataUint32ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kInt64ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kInt64ArraySetIndexed:
 | 
| -      return kTypedDataInt64ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kFloat32x4ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kFloat32x4ArraySetIndexed:
 | 
| -      return kTypedDataFloat32x4ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kInt32x4ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kInt32x4ArraySetIndexed:
 | 
| -      return kTypedDataInt32x4ArrayCid;
 | 
| -
 | 
| -    case MethodRecognizer::kFloat64x2ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kFloat64x2ArraySetIndexed:
 | 
| -      return kTypedDataFloat64x2ArrayCid;
 | 
| -
 | 
| -    default:
 | 
| -      break;
 | 
| -  }
 | 
| -  return kIllegalCid;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryReplaceWithStoreIndexed(InstanceCallInstr* call) {
 | 
| -  // Check for monomorphic IC data.
 | 
| -  if (!call->HasICData()) return false;
 | 
| -  const ICData& ic_data =
 | 
| -      ICData::Handle(I, call->ic_data()->AsUnaryClassChecks());
 | 
| -  if (ic_data.NumberOfChecks() != 1) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(ic_data.NumberOfUsedChecks() == 1);
 | 
| -  ASSERT(ic_data.HasOneTarget());
 | 
| -
 | 
| -  const Function& target = Function::Handle(I, ic_data.GetTargetAt(0));
 | 
| -  TargetEntryInstr* entry;
 | 
| -  Definition* last;
 | 
| -  if (!TryInlineRecognizedMethod(ic_data.GetReceiverClassIdAt(0),
 | 
| -                                 target,
 | 
| -                                 call,
 | 
| -                                 call->ArgumentAt(0),
 | 
| -                                 call->token_pos(),
 | 
| -                                 *call->ic_data(),
 | 
| -                                 &entry, &last)) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  // Insert receiver class check.
 | 
| -  AddReceiverCheck(call);
 | 
| -  // Remove the original push arguments.
 | 
| -  for (intptr_t i = 0; i < call->ArgumentCount(); ++i) {
 | 
| -    PushArgumentInstr* push = call->PushArgumentAt(i);
 | 
| -    push->ReplaceUsesWith(push->value()->definition());
 | 
| -    push->RemoveFromGraph();
 | 
| -  }
 | 
| -  // Replace all uses of this definition with the result.
 | 
| -  call->ReplaceUsesWith(last);
 | 
| -  // Finally insert the sequence other definition in place of this one in the
 | 
| -  // graph.
 | 
| -  call->previous()->LinkTo(entry->next());
 | 
| -  entry->UnuseAllInputs();  // Entry block is not in the graph.
 | 
| -  last->LinkTo(call);
 | 
| -  // Remove through the iterator.
 | 
| -  ASSERT(current_iterator()->Current() == call);
 | 
| -  current_iterator()->RemoveCurrentFromGraph();
 | 
| -  call->set_previous(NULL);
 | 
| -  call->set_next(NULL);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineSetIndexed(
 | 
| -    MethodRecognizer::Kind kind,
 | 
| -    const Function& target,
 | 
| -    Instruction* call,
 | 
| -    Definition* receiver,
 | 
| -    intptr_t token_pos,
 | 
| -    const ICData* ic_data,
 | 
| -    const ICData& value_check,
 | 
| -    TargetEntryInstr** entry,
 | 
| -    Definition** last) {
 | 
| -  intptr_t array_cid = MethodKindToCid(kind);
 | 
| -  ASSERT(array_cid != kIllegalCid);
 | 
| -
 | 
| -  Definition* array = receiver;
 | 
| -  Definition* index = call->ArgumentAt(1);
 | 
| -  Definition* stored_value = call->ArgumentAt(2);
 | 
| -
 | 
| -  *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(),
 | 
| -                                   call->GetBlock()->try_index());
 | 
| -  (*entry)->InheritDeoptTarget(I, call);
 | 
| -  Instruction* cursor = *entry;
 | 
| -  if (FLAG_enable_type_checks) {
 | 
| -    // Only type check for the value. A type check for the index is not
 | 
| -    // needed here because we insert a deoptimizing smi-check for the case
 | 
| -    // the index is not a smi.
 | 
| -    const AbstractType& value_type =
 | 
| -        AbstractType::ZoneHandle(I, target.ParameterTypeAt(2));
 | 
| -    Definition* instantiator = NULL;
 | 
| -    Definition* type_args = NULL;
 | 
| -    switch (array_cid) {
 | 
| -      case kArrayCid:
 | 
| -      case kGrowableObjectArrayCid: {
 | 
| -        const Class& instantiator_class =  Class::Handle(I, target.Owner());
 | 
| -        intptr_t type_arguments_field_offset =
 | 
| -            instantiator_class.type_arguments_field_offset();
 | 
| -        LoadFieldInstr* load_type_args =
 | 
| -            new(I) LoadFieldInstr(new(I) Value(array),
 | 
| -                                  type_arguments_field_offset,
 | 
| -                                  Type::ZoneHandle(I),  // No type.
 | 
| -                                  call->token_pos());
 | 
| -        cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                        load_type_args,
 | 
| -                                        NULL,
 | 
| -                                        FlowGraph::kValue);
 | 
| -
 | 
| -        instantiator = array;
 | 
| -        type_args = load_type_args;
 | 
| -        break;
 | 
| -      }
 | 
| -      case kTypedDataInt8ArrayCid:
 | 
| -      case kTypedDataUint8ArrayCid:
 | 
| -      case kTypedDataUint8ClampedArrayCid:
 | 
| -      case kExternalTypedDataUint8ArrayCid:
 | 
| -      case kExternalTypedDataUint8ClampedArrayCid:
 | 
| -      case kTypedDataInt16ArrayCid:
 | 
| -      case kTypedDataUint16ArrayCid:
 | 
| -      case kTypedDataInt32ArrayCid:
 | 
| -      case kTypedDataUint32ArrayCid:
 | 
| -      case kTypedDataInt64ArrayCid:
 | 
| -        ASSERT(value_type.IsIntType());
 | 
| -        // Fall through.
 | 
| -      case kTypedDataFloat32ArrayCid:
 | 
| -      case kTypedDataFloat64ArrayCid: {
 | 
| -        type_args = instantiator = flow_graph_->constant_null();
 | 
| -        ASSERT((array_cid != kTypedDataFloat32ArrayCid &&
 | 
| -                array_cid != kTypedDataFloat64ArrayCid) ||
 | 
| -               value_type.IsDoubleType());
 | 
| -        ASSERT(value_type.IsInstantiated());
 | 
| -        break;
 | 
| -      }
 | 
| -      case kTypedDataFloat32x4ArrayCid: {
 | 
| -        type_args = instantiator = flow_graph_->constant_null();
 | 
| -        ASSERT((array_cid != kTypedDataFloat32x4ArrayCid) ||
 | 
| -               value_type.IsFloat32x4Type());
 | 
| -        ASSERT(value_type.IsInstantiated());
 | 
| -        break;
 | 
| -      }
 | 
| -      case kTypedDataFloat64x2ArrayCid: {
 | 
| -        type_args = instantiator = flow_graph_->constant_null();
 | 
| -        ASSERT((array_cid != kTypedDataFloat64x2ArrayCid) ||
 | 
| -               value_type.IsFloat64x2Type());
 | 
| -        ASSERT(value_type.IsInstantiated());
 | 
| -        break;
 | 
| -      }
 | 
| -      default:
 | 
| -        // TODO(fschneider): Add support for other array types.
 | 
| -        UNREACHABLE();
 | 
| -    }
 | 
| -    AssertAssignableInstr* assert_value =
 | 
| -        new(I) AssertAssignableInstr(token_pos,
 | 
| -                                     new(I) Value(stored_value),
 | 
| -                                     new(I) Value(instantiator),
 | 
| -                                     new(I) Value(type_args),
 | 
| -                                     value_type,
 | 
| -                                     Symbols::Value(),
 | 
| -                                     call->deopt_id());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    assert_value,
 | 
| -                                    call->env(),
 | 
| -                                    FlowGraph::kValue);
 | 
| -  }
 | 
| -
 | 
| -  array_cid = PrepareInlineIndexedOp(call,
 | 
| -                                     array_cid,
 | 
| -                                     &array,
 | 
| -                                     index,
 | 
| -                                     &cursor);
 | 
| -
 | 
| -  // Check if store barrier is needed. Byte arrays don't need a store barrier.
 | 
| -  StoreBarrierType needs_store_barrier =
 | 
| -      (RawObject::IsTypedDataClassId(array_cid) ||
 | 
| -       RawObject::IsTypedDataViewClassId(array_cid) ||
 | 
| -       RawObject::IsExternalTypedDataClassId(array_cid)) ? kNoStoreBarrier
 | 
| -                                                         : kEmitStoreBarrier;
 | 
| -
 | 
| -  // No need to class check stores to Int32 and Uint32 arrays because
 | 
| -  // we insert unboxing instructions below which include a class check.
 | 
| -  if ((array_cid != kTypedDataUint32ArrayCid) &&
 | 
| -      (array_cid != kTypedDataInt32ArrayCid) &&
 | 
| -      !value_check.IsNull()) {
 | 
| -    // No store barrier needed because checked value is a smi, an unboxed mint,
 | 
| -    // an unboxed double, an unboxed Float32x4, or unboxed Int32x4.
 | 
| -    needs_store_barrier = kNoStoreBarrier;
 | 
| -    Instruction* check = GetCheckClass(
 | 
| -        stored_value, value_check, call->deopt_id(), call->token_pos());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    check,
 | 
| -                                    call->env(),
 | 
| -                                    FlowGraph::kEffect);
 | 
| -  }
 | 
| -
 | 
| -  if (array_cid == kTypedDataFloat32ArrayCid) {
 | 
| -    stored_value =
 | 
| -        new(I) DoubleToFloatInstr(
 | 
| -            new(I) Value(stored_value), call->deopt_id());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    stored_value,
 | 
| -                                    NULL,
 | 
| -                                    FlowGraph::kValue);
 | 
| -  } else if (array_cid == kTypedDataInt32ArrayCid) {
 | 
| -    stored_value = new(I) UnboxInt32Instr(
 | 
| -        UnboxInt32Instr::kTruncate,
 | 
| -        new(I) Value(stored_value),
 | 
| -        call->deopt_id());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    stored_value,
 | 
| -                                    call->env(),
 | 
| -                                    FlowGraph::kValue);
 | 
| -  } else if (array_cid == kTypedDataUint32ArrayCid) {
 | 
| -    stored_value = new(I) UnboxUint32Instr(
 | 
| -        new(I) Value(stored_value),
 | 
| -        call->deopt_id());
 | 
| -    ASSERT(stored_value->AsUnboxInteger()->is_truncating());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    stored_value,
 | 
| -                                    call->env(),
 | 
| -                                    FlowGraph::kValue);
 | 
| -  }
 | 
| -
 | 
| -  const intptr_t index_scale = Instance::ElementSizeFor(array_cid);
 | 
| -  *last = new(I) StoreIndexedInstr(new(I) Value(array),
 | 
| -                                   new(I) Value(index),
 | 
| -                                   new(I) Value(stored_value),
 | 
| -                                   needs_store_barrier,
 | 
| -                                   index_scale,
 | 
| -                                   array_cid,
 | 
| -                                   call->deopt_id(),
 | 
| -                                   call->token_pos());
 | 
| -  flow_graph()->AppendTo(cursor,
 | 
| -                         *last,
 | 
| -                         call->env(),
 | 
| -                         FlowGraph::kEffect);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineRecognizedMethod(intptr_t receiver_cid,
 | 
| -                                                   const Function& target,
 | 
| -                                                   Instruction* call,
 | 
| -                                                   Definition* receiver,
 | 
| -                                                   intptr_t token_pos,
 | 
| -                                                   const ICData& ic_data,
 | 
| -                                                   TargetEntryInstr** entry,
 | 
| -                                                   Definition** last) {
 | 
| -  ICData& value_check = ICData::ZoneHandle(I);
 | 
| -  MethodRecognizer::Kind kind = MethodRecognizer::RecognizeKind(target);
 | 
| -  switch (kind) {
 | 
| -    // Recognized [] operators.
 | 
| -    case MethodRecognizer::kImmutableArrayGetIndexed:
 | 
| -    case MethodRecognizer::kObjectArrayGetIndexed:
 | 
| -    case MethodRecognizer::kGrowableArrayGetIndexed:
 | 
| -    case MethodRecognizer::kInt8ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint8ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint8ClampedArrayGetIndexed:
 | 
| -    case MethodRecognizer::kExternalUint8ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kExternalUint8ClampedArrayGetIndexed:
 | 
| -    case MethodRecognizer::kInt16ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint16ArrayGetIndexed:
 | 
| -      return InlineGetIndexed(kind, call, receiver, ic_data, entry, last);
 | 
| -    case MethodRecognizer::kFloat32ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kFloat64ArrayGetIndexed:
 | 
| -      if (!CanUnboxDouble()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineGetIndexed(kind, call, receiver, ic_data, entry, last);
 | 
| -    case MethodRecognizer::kFloat32x4ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kFloat64x2ArrayGetIndexed:
 | 
| -      if (!ShouldInlineSimd()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineGetIndexed(kind, call, receiver, ic_data, entry, last);
 | 
| -    case MethodRecognizer::kInt32ArrayGetIndexed:
 | 
| -    case MethodRecognizer::kUint32ArrayGetIndexed:
 | 
| -      if (!CanUnboxInt32()) return false;
 | 
| -      return InlineGetIndexed(kind, call, receiver, ic_data, entry, last);
 | 
| -
 | 
| -    case MethodRecognizer::kInt64ArrayGetIndexed:
 | 
| -      if (!ShouldInlineInt64ArrayOps()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineGetIndexed(kind, call, receiver, ic_data, entry, last);
 | 
| -    // Recognized []= operators.
 | 
| -    case MethodRecognizer::kObjectArraySetIndexed:
 | 
| -    case MethodRecognizer::kGrowableArraySetIndexed:
 | 
| -      if (ArgIsAlways(kSmiCid, ic_data, 2)) {
 | 
| -        value_check = ic_data.AsUnaryClassChecksForArgNr(2);
 | 
| -      }
 | 
| -      return InlineSetIndexed(kind, target, call, receiver, token_pos,
 | 
| -                              &ic_data, value_check, entry, last);
 | 
| -    case MethodRecognizer::kInt8ArraySetIndexed:
 | 
| -    case MethodRecognizer::kUint8ArraySetIndexed:
 | 
| -    case MethodRecognizer::kUint8ClampedArraySetIndexed:
 | 
| -    case MethodRecognizer::kExternalUint8ArraySetIndexed:
 | 
| -    case MethodRecognizer::kExternalUint8ClampedArraySetIndexed:
 | 
| -    case MethodRecognizer::kInt16ArraySetIndexed:
 | 
| -    case MethodRecognizer::kUint16ArraySetIndexed:
 | 
| -      if (!ArgIsAlways(kSmiCid, ic_data, 2)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      value_check = ic_data.AsUnaryClassChecksForArgNr(2);
 | 
| -      return InlineSetIndexed(kind, target, call, receiver, token_pos,
 | 
| -                              &ic_data, value_check, entry, last);
 | 
| -    case MethodRecognizer::kInt32ArraySetIndexed:
 | 
| -    case MethodRecognizer::kUint32ArraySetIndexed:
 | 
| -      // Check that value is always smi or mint. We use Int32/Uint32 unboxing
 | 
| -      // which can only deal unbox these values.
 | 
| -      value_check = ic_data.AsUnaryClassChecksForArgNr(2);
 | 
| -      if (!HasOnlySmiOrMint(value_check)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineSetIndexed(kind, target, call, receiver, token_pos,
 | 
| -                              &ic_data, value_check, entry, last);
 | 
| -    case MethodRecognizer::kInt64ArraySetIndexed:
 | 
| -      if (!ShouldInlineInt64ArrayOps()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineSetIndexed(kind, target, call, receiver, token_pos,
 | 
| -                              &ic_data, value_check, entry, last);
 | 
| -    case MethodRecognizer::kFloat32ArraySetIndexed:
 | 
| -    case MethodRecognizer::kFloat64ArraySetIndexed:
 | 
| -      if (!CanUnboxDouble()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      // Check that value is always double.
 | 
| -      if (!ArgIsAlways(kDoubleCid, ic_data, 2)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      value_check = ic_data.AsUnaryClassChecksForArgNr(2);
 | 
| -      return InlineSetIndexed(kind, target, call, receiver, token_pos,
 | 
| -                              &ic_data, value_check, entry, last);
 | 
| -    case MethodRecognizer::kFloat32x4ArraySetIndexed:
 | 
| -      if (!ShouldInlineSimd()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      // Check that value is always a Float32x4.
 | 
| -      if (!ArgIsAlways(kFloat32x4Cid, ic_data, 2)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      value_check = ic_data.AsUnaryClassChecksForArgNr(2);
 | 
| -      return InlineSetIndexed(kind, target, call, receiver, token_pos,
 | 
| -                              &ic_data, value_check, entry, last);
 | 
| -    case MethodRecognizer::kFloat64x2ArraySetIndexed:
 | 
| -      if (!ShouldInlineSimd()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      // Check that value is always a Float32x4.
 | 
| -      if (!ArgIsAlways(kFloat64x2Cid, ic_data, 2)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      value_check = ic_data.AsUnaryClassChecksForArgNr(2);
 | 
| -      return InlineSetIndexed(kind, target, call, receiver, token_pos,
 | 
| -                              &ic_data, value_check, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetInt8:
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataInt8ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetUint8:
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataUint8ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetInt16:
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataInt16ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetUint16:
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataUint16ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetInt32:
 | 
| -      if (!CanUnboxInt32()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataInt32ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetUint32:
 | 
| -      if (!CanUnboxInt32()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataUint32ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetFloat32:
 | 
| -      if (!CanUnboxDouble()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataFloat32ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetFloat64:
 | 
| -      if (!CanUnboxDouble()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataFloat64ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetFloat32x4:
 | 
| -      if (!ShouldInlineSimd()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataFloat32x4ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseGetInt32x4:
 | 
| -      if (!ShouldInlineSimd()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewLoad(call, receiver, receiver_cid,
 | 
| -                                     kTypedDataInt32x4ArrayCid,
 | 
| -                                     ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetInt8:
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataInt8ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetUint8:
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataUint8ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetInt16:
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataInt16ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetUint16:
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataUint16ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetInt32:
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataInt32ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetUint32:
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataUint32ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetFloat32:
 | 
| -      if (!CanUnboxDouble()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataFloat32ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetFloat64:
 | 
| -      if (!CanUnboxDouble()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataFloat64ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetFloat32x4:
 | 
| -      if (!ShouldInlineSimd()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataFloat32x4ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kByteArrayBaseSetInt32x4:
 | 
| -      if (!ShouldInlineSimd()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      return InlineByteArrayViewStore(target, call, receiver, receiver_cid,
 | 
| -                                      kTypedDataInt32x4ArrayCid,
 | 
| -                                      ic_data, entry, last);
 | 
| -    case MethodRecognizer::kStringBaseCodeUnitAt:
 | 
| -      return InlineStringCodeUnitAt(call, receiver_cid, entry, last);
 | 
| -    case MethodRecognizer::kStringBaseCharAt:
 | 
| -      return InlineStringBaseCharAt(call, receiver_cid, entry, last);
 | 
| -    case MethodRecognizer::kDoubleAdd:
 | 
| -      return InlineDoubleOp(Token::kADD, call, entry, last);
 | 
| -    case MethodRecognizer::kDoubleSub:
 | 
| -      return InlineDoubleOp(Token::kSUB, call, entry, last);
 | 
| -    case MethodRecognizer::kDoubleMul:
 | 
| -      return InlineDoubleOp(Token::kMUL, call, entry, last);
 | 
| -    case MethodRecognizer::kDoubleDiv:
 | 
| -      return InlineDoubleOp(Token::kDIV, call, entry, last);
 | 
| -    default:
 | 
| -      return false;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -intptr_t FlowGraphOptimizer::PrepareInlineIndexedOp(Instruction* call,
 | 
| -                                                    intptr_t array_cid,
 | 
| -                                                    Definition** array,
 | 
| -                                                    Definition* index,
 | 
| -                                                    Instruction** cursor) {
 | 
| -  // Insert index smi check.
 | 
| -  *cursor = flow_graph()->AppendTo(
 | 
| -      *cursor,
 | 
| -      new(I) CheckSmiInstr(new(I) Value(index),
 | 
| -                           call->deopt_id(),
 | 
| -                           call->token_pos()),
 | 
| -      call->env(),
 | 
| -      FlowGraph::kEffect);
 | 
| -
 | 
| -  // Insert array length load and bounds check.
 | 
| -  LoadFieldInstr* length =
 | 
| -      new(I) LoadFieldInstr(
 | 
| -          new(I) Value(*array),
 | 
| -          CheckArrayBoundInstr::LengthOffsetFor(array_cid),
 | 
| -          Type::ZoneHandle(I, Type::SmiType()),
 | 
| -          call->token_pos());
 | 
| -  length->set_is_immutable(
 | 
| -      CheckArrayBoundInstr::IsFixedLengthArrayType(array_cid));
 | 
| -  length->set_result_cid(kSmiCid);
 | 
| -  length->set_recognized_kind(
 | 
| -      LoadFieldInstr::RecognizedKindFromArrayCid(array_cid));
 | 
| -  *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                   length,
 | 
| -                                   NULL,
 | 
| -                                   FlowGraph::kValue);
 | 
| -
 | 
| -  *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                   new(I) CheckArrayBoundInstr(
 | 
| -                                       new(I) Value(length),
 | 
| -                                       new(I) Value(index),
 | 
| -                                       call->deopt_id()),
 | 
| -                                   call->env(),
 | 
| -                                   FlowGraph::kEffect);
 | 
| -
 | 
| -  if (array_cid == kGrowableObjectArrayCid) {
 | 
| -    // Insert data elements load.
 | 
| -    LoadFieldInstr* elements =
 | 
| -        new(I) LoadFieldInstr(
 | 
| -            new(I) Value(*array),
 | 
| -            GrowableObjectArray::data_offset(),
 | 
| -            Type::ZoneHandle(I, Type::DynamicType()),
 | 
| -            call->token_pos());
 | 
| -    elements->set_result_cid(kArrayCid);
 | 
| -    *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                     elements,
 | 
| -                                     NULL,
 | 
| -                                     FlowGraph::kValue);
 | 
| -    // Load from the data from backing store which is a fixed-length array.
 | 
| -    *array = elements;
 | 
| -    array_cid = kArrayCid;
 | 
| -  } else if (RawObject::IsExternalTypedDataClassId(array_cid)) {
 | 
| -    LoadUntaggedInstr* elements =
 | 
| -        new(I) LoadUntaggedInstr(new(I) Value(*array),
 | 
| -                                 ExternalTypedData::data_offset());
 | 
| -    *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                     elements,
 | 
| -                                     NULL,
 | 
| -                                     FlowGraph::kValue);
 | 
| -    *array = elements;
 | 
| -  }
 | 
| -  return array_cid;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineGetIndexed(MethodRecognizer::Kind kind,
 | 
| -                                          Instruction* call,
 | 
| -                                          Definition* receiver,
 | 
| -                                          const ICData& ic_data,
 | 
| -                                          TargetEntryInstr** entry,
 | 
| -                                          Definition** last) {
 | 
| -  intptr_t array_cid = MethodKindToCid(kind);
 | 
| -  ASSERT(array_cid != kIllegalCid);
 | 
| -
 | 
| -  Definition* array = receiver;
 | 
| -  Definition* index = call->ArgumentAt(1);
 | 
| -  *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(),
 | 
| -                                   call->GetBlock()->try_index());
 | 
| -  (*entry)->InheritDeoptTarget(I, call);
 | 
| -  Instruction* cursor = *entry;
 | 
| -
 | 
| -  array_cid = PrepareInlineIndexedOp(call,
 | 
| -                                     array_cid,
 | 
| -                                     &array,
 | 
| -                                     index,
 | 
| -                                     &cursor);
 | 
| -
 | 
| -  intptr_t deopt_id = Isolate::kNoDeoptId;
 | 
| -  if ((array_cid == kTypedDataInt32ArrayCid) ||
 | 
| -      (array_cid == kTypedDataUint32ArrayCid)) {
 | 
| -    // Deoptimization may be needed if result does not always fit in a Smi.
 | 
| -    deopt_id = (kSmiBits >= 32) ? Isolate::kNoDeoptId : call->deopt_id();
 | 
| -  }
 | 
| -
 | 
| -  // Array load and return.
 | 
| -  intptr_t index_scale = Instance::ElementSizeFor(array_cid);
 | 
| -  *last = new(I) LoadIndexedInstr(new(I) Value(array),
 | 
| -                                  new(I) Value(index),
 | 
| -                                  index_scale,
 | 
| -                                  array_cid,
 | 
| -                                  deopt_id,
 | 
| -                                  call->token_pos());
 | 
| -  cursor = flow_graph()->AppendTo(
 | 
| -      cursor,
 | 
| -      *last,
 | 
| -      deopt_id != Isolate::kNoDeoptId ? call->env() : NULL,
 | 
| -      FlowGraph::kValue);
 | 
| -
 | 
| -  if (array_cid == kTypedDataFloat32ArrayCid) {
 | 
| -    *last = new(I) FloatToDoubleInstr(new(I) Value(*last), deopt_id);
 | 
| -    flow_graph()->AppendTo(cursor,
 | 
| -                           *last,
 | 
| -                           deopt_id != Isolate::kNoDeoptId ? call->env() : NULL,
 | 
| -                           FlowGraph::kValue);
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryReplaceWithLoadIndexed(InstanceCallInstr* call) {
 | 
| -  // Check for monomorphic IC data.
 | 
| -  if (!call->HasICData()) return false;
 | 
| -  const ICData& ic_data =
 | 
| -      ICData::Handle(I, call->ic_data()->AsUnaryClassChecks());
 | 
| -  if (ic_data.NumberOfChecks() != 1) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(ic_data.NumberOfUsedChecks() == 1);
 | 
| -  ASSERT(ic_data.HasOneTarget());
 | 
| -
 | 
| -  const Function& target = Function::Handle(I, ic_data.GetTargetAt(0));
 | 
| -  TargetEntryInstr* entry;
 | 
| -  Definition* last;
 | 
| -  if (!TryInlineRecognizedMethod(ic_data.GetReceiverClassIdAt(0),
 | 
| -                                 target,
 | 
| -                                 call,
 | 
| -                                 call->ArgumentAt(0),
 | 
| -                                 call->token_pos(),
 | 
| -                                 *call->ic_data(),
 | 
| -                                 &entry, &last)) {
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  // Insert receiver class check.
 | 
| -  AddReceiverCheck(call);
 | 
| -  // Remove the original push arguments.
 | 
| -  for (intptr_t i = 0; i < call->ArgumentCount(); ++i) {
 | 
| -    PushArgumentInstr* push = call->PushArgumentAt(i);
 | 
| -    push->ReplaceUsesWith(push->value()->definition());
 | 
| -    push->RemoveFromGraph();
 | 
| -  }
 | 
| -  // Replace all uses of this definition with the result.
 | 
| -  call->ReplaceUsesWith(last);
 | 
| -  // Finally insert the sequence other definition in place of this one in the
 | 
| -  // graph.
 | 
| -  call->previous()->LinkTo(entry->next());
 | 
| -  entry->UnuseAllInputs();  // Entry block is not in the graph.
 | 
| -  last->LinkTo(call);
 | 
| -  // Remove through the iterator.
 | 
| -  ASSERT(current_iterator()->Current() == call);
 | 
| -  current_iterator()->RemoveCurrentFromGraph();
 | 
| -  call->set_previous(NULL);
 | 
| -  call->set_next(NULL);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Return true if d is a string of length one (a constant or result from
 | 
| -// from string-from-char-code instruction.
 | 
| -static bool IsLengthOneString(Definition* d) {
 | 
| -  if (d->IsConstant()) {
 | 
| -    const Object& obj = d->AsConstant()->value();
 | 
| -    if (obj.IsString()) {
 | 
| -      return String::Cast(obj).Length() == 1;
 | 
| -    } else {
 | 
| -      return false;
 | 
| -    }
 | 
| -  } else {
 | 
| -    return d->IsStringFromCharCode();
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Returns true if the string comparison was converted into char-code
 | 
| -// comparison. Conversion is only possible for strings of length one.
 | 
| -// E.g., detect str[x] == "x"; and use an integer comparison of char-codes.
 | 
| -// TODO(srdjan): Expand for two-byte and external strings.
 | 
| -bool FlowGraphOptimizer::TryStringLengthOneEquality(InstanceCallInstr* call,
 | 
| -                                                    Token::Kind op_kind) {
 | 
| -  ASSERT(HasOnlyTwoOf(*call->ic_data(), kOneByteStringCid));
 | 
| -  // Check that left and right are length one strings (either string constants
 | 
| -  // or results of string-from-char-code.
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -  Value* left_val = NULL;
 | 
| -  Definition* to_remove_left = NULL;
 | 
| -  if (IsLengthOneString(right)) {
 | 
| -    // Swap, since we know that both arguments are strings
 | 
| -    Definition* temp = left;
 | 
| -    left = right;
 | 
| -    right = temp;
 | 
| -  }
 | 
| -  if (IsLengthOneString(left)) {
 | 
| -    // Optimize if left is a string with length one (either constant or
 | 
| -    // result of string-from-char-code.
 | 
| -    if (left->IsConstant()) {
 | 
| -      ConstantInstr* left_const = left->AsConstant();
 | 
| -      const String& str = String::Cast(left_const->value());
 | 
| -      ASSERT(str.Length() == 1);
 | 
| -      ConstantInstr* char_code_left = flow_graph()->GetConstant(
 | 
| -          Smi::ZoneHandle(I, Smi::New(static_cast<intptr_t>(str.CharAt(0)))));
 | 
| -      left_val = new(I) Value(char_code_left);
 | 
| -    } else if (left->IsStringFromCharCode()) {
 | 
| -      // Use input of string-from-charcode as left value.
 | 
| -      StringFromCharCodeInstr* instr = left->AsStringFromCharCode();
 | 
| -      left_val = new(I) Value(instr->char_code()->definition());
 | 
| -      to_remove_left = instr;
 | 
| -    } else {
 | 
| -      // IsLengthOneString(left) should have been false.
 | 
| -      UNREACHABLE();
 | 
| -    }
 | 
| -
 | 
| -    Definition* to_remove_right = NULL;
 | 
| -    Value* right_val = NULL;
 | 
| -    if (right->IsStringFromCharCode()) {
 | 
| -      // Skip string-from-char-code, and use its input as right value.
 | 
| -      StringFromCharCodeInstr* right_instr = right->AsStringFromCharCode();
 | 
| -      right_val = new(I) Value(right_instr->char_code()->definition());
 | 
| -      to_remove_right = right_instr;
 | 
| -    } else {
 | 
| -      const ICData& unary_checks_1 =
 | 
| -          ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecksForArgNr(1));
 | 
| -      AddCheckClass(right,
 | 
| -                    unary_checks_1,
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      // String-to-char-code instructions returns -1 (illegal charcode) if
 | 
| -      // string is not of length one.
 | 
| -      StringToCharCodeInstr* char_code_right =
 | 
| -          new(I) StringToCharCodeInstr(new(I) Value(right), kOneByteStringCid);
 | 
| -      InsertBefore(call, char_code_right, call->env(), FlowGraph::kValue);
 | 
| -      right_val = new(I) Value(char_code_right);
 | 
| -    }
 | 
| -
 | 
| -    // Comparing char-codes instead of strings.
 | 
| -    EqualityCompareInstr* comp =
 | 
| -        new(I) EqualityCompareInstr(call->token_pos(),
 | 
| -                                    op_kind,
 | 
| -                                    left_val,
 | 
| -                                    right_val,
 | 
| -                                    kSmiCid,
 | 
| -                                    call->deopt_id());
 | 
| -    ReplaceCall(call, comp);
 | 
| -
 | 
| -    // Remove dead instructions.
 | 
| -    if ((to_remove_left != NULL) &&
 | 
| -        (to_remove_left->input_use_list() == NULL)) {
 | 
| -      to_remove_left->ReplaceUsesWith(flow_graph()->constant_null());
 | 
| -      to_remove_left->RemoveFromGraph();
 | 
| -    }
 | 
| -    if ((to_remove_right != NULL) &&
 | 
| -        (to_remove_right->input_use_list() == NULL)) {
 | 
| -      to_remove_right->ReplaceUsesWith(flow_graph()->constant_null());
 | 
| -      to_remove_right->RemoveFromGraph();
 | 
| -    }
 | 
| -    return true;
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool SmiFitsInDouble() { return kSmiBits < 53; }
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryReplaceWithEqualityOp(InstanceCallInstr* call,
 | 
| -                                                  Token::Kind op_kind) {
 | 
| -  const ICData& ic_data = *call->ic_data();
 | 
| -  ASSERT(ic_data.NumArgsTested() == 2);
 | 
| -
 | 
| -  ASSERT(call->ArgumentCount() == 2);
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -
 | 
| -  intptr_t cid = kIllegalCid;
 | 
| -  if (HasOnlyTwoOf(ic_data, kOneByteStringCid)) {
 | 
| -    if (TryStringLengthOneEquality(call, op_kind)) {
 | 
| -      return true;
 | 
| -    } else {
 | 
| -      return false;
 | 
| -    }
 | 
| -  } else if (HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -    InsertBefore(call,
 | 
| -                 new(I) CheckSmiInstr(new(I) Value(left),
 | 
| -                                      call->deopt_id(),
 | 
| -                                      call->token_pos()),
 | 
| -                 call->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -    InsertBefore(call,
 | 
| -                 new(I) CheckSmiInstr(new(I) Value(right),
 | 
| -                                      call->deopt_id(),
 | 
| -                                      call->token_pos()),
 | 
| -                 call->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -    cid = kSmiCid;
 | 
| -  } else if (HasTwoMintOrSmi(ic_data) &&
 | 
| -             FlowGraphCompiler::SupportsUnboxedMints()) {
 | 
| -    cid = kMintCid;
 | 
| -  } else if (HasTwoDoubleOrSmi(ic_data) && CanUnboxDouble()) {
 | 
| -    // Use double comparison.
 | 
| -    if (SmiFitsInDouble()) {
 | 
| -      cid = kDoubleCid;
 | 
| -    } else {
 | 
| -      if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) {
 | 
| -        // We cannot use double comparison on two smis. Need polymorphic
 | 
| -        // call.
 | 
| -        return false;
 | 
| -      } else {
 | 
| -        InsertBefore(call,
 | 
| -                     new(I) CheckEitherNonSmiInstr(
 | 
| -                         new(I) Value(left),
 | 
| -                         new(I) Value(right),
 | 
| -                         call->deopt_id()),
 | 
| -                     call->env(),
 | 
| -                     FlowGraph::kEffect);
 | 
| -        cid = kDoubleCid;
 | 
| -      }
 | 
| -    }
 | 
| -  } else {
 | 
| -    // Check if ICDData contains checks with Smi/Null combinations. In that case
 | 
| -    // we can still emit the optimized Smi equality operation but need to add
 | 
| -    // checks for null or Smi.
 | 
| -    GrowableArray<intptr_t> smi_or_null(2);
 | 
| -    smi_or_null.Add(kSmiCid);
 | 
| -    smi_or_null.Add(kNullCid);
 | 
| -    if (ICDataHasOnlyReceiverArgumentClassIds(ic_data,
 | 
| -                                              smi_or_null,
 | 
| -                                              smi_or_null)) {
 | 
| -      const ICData& unary_checks_0 =
 | 
| -          ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks());
 | 
| -      AddCheckClass(left,
 | 
| -                    unary_checks_0,
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -
 | 
| -      const ICData& unary_checks_1 =
 | 
| -          ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecksForArgNr(1));
 | 
| -      AddCheckClass(right,
 | 
| -                    unary_checks_1,
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      cid = kSmiCid;
 | 
| -    } else {
 | 
| -      // Shortcut for equality with null.
 | 
| -      ConstantInstr* right_const = right->AsConstant();
 | 
| -      ConstantInstr* left_const = left->AsConstant();
 | 
| -      if ((right_const != NULL && right_const->value().IsNull()) ||
 | 
| -          (left_const != NULL && left_const->value().IsNull())) {
 | 
| -        StrictCompareInstr* comp =
 | 
| -            new(I) StrictCompareInstr(call->token_pos(),
 | 
| -                                      Token::kEQ_STRICT,
 | 
| -                                      new(I) Value(left),
 | 
| -                                      new(I) Value(right),
 | 
| -                                      false);  // No number check.
 | 
| -        ReplaceCall(call, comp);
 | 
| -        return true;
 | 
| -      }
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -  ASSERT(cid != kIllegalCid);
 | 
| -  EqualityCompareInstr* comp = new(I) EqualityCompareInstr(call->token_pos(),
 | 
| -                                                           op_kind,
 | 
| -                                                           new(I) Value(left),
 | 
| -                                                           new(I) Value(right),
 | 
| -                                                           cid,
 | 
| -                                                           call->deopt_id());
 | 
| -  ReplaceCall(call, comp);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryReplaceWithRelationalOp(InstanceCallInstr* call,
 | 
| -                                                    Token::Kind op_kind) {
 | 
| -  const ICData& ic_data = *call->ic_data();
 | 
| -  ASSERT(ic_data.NumArgsTested() == 2);
 | 
| -
 | 
| -  ASSERT(call->ArgumentCount() == 2);
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -
 | 
| -  intptr_t cid = kIllegalCid;
 | 
| -  if (HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -    InsertBefore(call,
 | 
| -                 new(I) CheckSmiInstr(new(I) Value(left),
 | 
| -                                      call->deopt_id(),
 | 
| -                                      call->token_pos()),
 | 
| -                 call->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -    InsertBefore(call,
 | 
| -                 new(I) CheckSmiInstr(new(I) Value(right),
 | 
| -                                      call->deopt_id(),
 | 
| -                                      call->token_pos()),
 | 
| -                 call->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -    cid = kSmiCid;
 | 
| -  } else if (HasTwoMintOrSmi(ic_data) &&
 | 
| -             FlowGraphCompiler::SupportsUnboxedMints()) {
 | 
| -    cid = kMintCid;
 | 
| -  } else if (HasTwoDoubleOrSmi(ic_data) && CanUnboxDouble()) {
 | 
| -    // Use double comparison.
 | 
| -    if (SmiFitsInDouble()) {
 | 
| -      cid = kDoubleCid;
 | 
| -    } else {
 | 
| -      if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) {
 | 
| -        // We cannot use double comparison on two smis. Need polymorphic
 | 
| -        // call.
 | 
| -        return false;
 | 
| -      } else {
 | 
| -        InsertBefore(call,
 | 
| -                     new(I) CheckEitherNonSmiInstr(
 | 
| -                         new(I) Value(left),
 | 
| -                         new(I) Value(right),
 | 
| -                         call->deopt_id()),
 | 
| -                     call->env(),
 | 
| -                     FlowGraph::kEffect);
 | 
| -        cid = kDoubleCid;
 | 
| -      }
 | 
| -    }
 | 
| -  } else {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(cid != kIllegalCid);
 | 
| -  RelationalOpInstr* comp = new(I) RelationalOpInstr(call->token_pos(),
 | 
| -                                                     op_kind,
 | 
| -                                                     new(I) Value(left),
 | 
| -                                                     new(I) Value(right),
 | 
| -                                                     cid,
 | 
| -                                                     call->deopt_id());
 | 
| -  ReplaceCall(call, comp);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryReplaceWithBinaryOp(InstanceCallInstr* call,
 | 
| -                                                Token::Kind op_kind) {
 | 
| -  intptr_t operands_type = kIllegalCid;
 | 
| -  ASSERT(call->HasICData());
 | 
| -  const ICData& ic_data = *call->ic_data();
 | 
| -  switch (op_kind) {
 | 
| -    case Token::kADD:
 | 
| -    case Token::kSUB:
 | 
| -    case Token::kMUL:
 | 
| -      if (HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -        // Don't generate smi code if the IC data is marked because
 | 
| -        // of an overflow.
 | 
| -        operands_type = ic_data.HasDeoptReason(ICData::kDeoptBinarySmiOp)
 | 
| -            ? kMintCid
 | 
| -            : kSmiCid;
 | 
| -      } else if (HasTwoMintOrSmi(ic_data) &&
 | 
| -                 FlowGraphCompiler::SupportsUnboxedMints()) {
 | 
| -        // Don't generate mint code if the IC data is marked because of an
 | 
| -        // overflow.
 | 
| -        if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) return false;
 | 
| -        operands_type = kMintCid;
 | 
| -      } else if (ShouldSpecializeForDouble(ic_data)) {
 | 
| -        operands_type = kDoubleCid;
 | 
| -      } else if (HasOnlyTwoOf(ic_data, kFloat32x4Cid)) {
 | 
| -        operands_type = kFloat32x4Cid;
 | 
| -      } else if (HasOnlyTwoOf(ic_data, kInt32x4Cid)) {
 | 
| -        ASSERT(op_kind != Token::kMUL);  // Int32x4 doesn't have a multiply op.
 | 
| -        operands_type = kInt32x4Cid;
 | 
| -      } else if (HasOnlyTwoOf(ic_data, kFloat64x2Cid)) {
 | 
| -        operands_type = kFloat64x2Cid;
 | 
| -      } else {
 | 
| -        return false;
 | 
| -      }
 | 
| -      break;
 | 
| -    case Token::kDIV:
 | 
| -      if (ShouldSpecializeForDouble(ic_data) ||
 | 
| -          HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -        operands_type = kDoubleCid;
 | 
| -      } else if (HasOnlyTwoOf(ic_data, kFloat32x4Cid)) {
 | 
| -        operands_type = kFloat32x4Cid;
 | 
| -      } else if (HasOnlyTwoOf(ic_data, kFloat64x2Cid)) {
 | 
| -        operands_type = kFloat64x2Cid;
 | 
| -      } else {
 | 
| -        return false;
 | 
| -      }
 | 
| -      break;
 | 
| -    case Token::kBIT_AND:
 | 
| -    case Token::kBIT_OR:
 | 
| -    case Token::kBIT_XOR:
 | 
| -      if (HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -        operands_type = kSmiCid;
 | 
| -      } else if (HasTwoMintOrSmi(ic_data)) {
 | 
| -        operands_type = kMintCid;
 | 
| -      } else if (HasOnlyTwoOf(ic_data, kInt32x4Cid)) {
 | 
| -        operands_type = kInt32x4Cid;
 | 
| -      } else {
 | 
| -        return false;
 | 
| -      }
 | 
| -      break;
 | 
| -    case Token::kSHR:
 | 
| -    case Token::kSHL:
 | 
| -      if (HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -        // Left shift may overflow from smi into mint or big ints.
 | 
| -        // Don't generate smi code if the IC data is marked because
 | 
| -        // of an overflow.
 | 
| -        if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) {
 | 
| -          return false;
 | 
| -        }
 | 
| -        operands_type = ic_data.HasDeoptReason(ICData::kDeoptBinarySmiOp)
 | 
| -            ? kMintCid
 | 
| -            : kSmiCid;
 | 
| -      } else if (HasTwoMintOrSmi(ic_data) &&
 | 
| -                 HasOnlyOneSmi(ICData::Handle(I,
 | 
| -                     ic_data.AsUnaryClassChecksForArgNr(1)))) {
 | 
| -        // Don't generate mint code if the IC data is marked because of an
 | 
| -        // overflow.
 | 
| -        if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) {
 | 
| -          return false;
 | 
| -        }
 | 
| -        // Check for smi/mint << smi or smi/mint >> smi.
 | 
| -        operands_type = kMintCid;
 | 
| -      } else {
 | 
| -        return false;
 | 
| -      }
 | 
| -      break;
 | 
| -    case Token::kMOD:
 | 
| -    case Token::kTRUNCDIV:
 | 
| -      if (HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -        if (ic_data.HasDeoptReason(ICData::kDeoptBinarySmiOp)) {
 | 
| -          return false;
 | 
| -        }
 | 
| -        operands_type = kSmiCid;
 | 
| -      } else {
 | 
| -        return false;
 | 
| -      }
 | 
| -      break;
 | 
| -    default:
 | 
| -      UNREACHABLE();
 | 
| -  }
 | 
| -
 | 
| -  ASSERT(call->ArgumentCount() == 2);
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -  if (operands_type == kDoubleCid) {
 | 
| -    if (!CanUnboxDouble()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -    // Check that either left or right are not a smi.  Result of a
 | 
| -    // binary operation with two smis is a smi not a double, except '/' which
 | 
| -    // returns a double for two smis.
 | 
| -    if (op_kind != Token::kDIV) {
 | 
| -      InsertBefore(call,
 | 
| -                   new(I) CheckEitherNonSmiInstr(
 | 
| -                       new(I) Value(left),
 | 
| -                       new(I) Value(right),
 | 
| -                       call->deopt_id()),
 | 
| -                   call->env(),
 | 
| -                   FlowGraph::kEffect);
 | 
| -    }
 | 
| -
 | 
| -    BinaryDoubleOpInstr* double_bin_op =
 | 
| -        new(I) BinaryDoubleOpInstr(op_kind,
 | 
| -                                   new(I) Value(left),
 | 
| -                                   new(I) Value(right),
 | 
| -                                   call->deopt_id(), call->token_pos());
 | 
| -    ReplaceCall(call, double_bin_op);
 | 
| -  } else if (operands_type == kMintCid) {
 | 
| -    if (!FlowGraphCompiler::SupportsUnboxedMints()) return false;
 | 
| -    if ((op_kind == Token::kSHR) || (op_kind == Token::kSHL)) {
 | 
| -      ShiftMintOpInstr* shift_op =
 | 
| -          new(I) ShiftMintOpInstr(
 | 
| -              op_kind, new(I) Value(left), new(I) Value(right),
 | 
| -              call->deopt_id());
 | 
| -      ReplaceCall(call, shift_op);
 | 
| -    } else {
 | 
| -      BinaryMintOpInstr* bin_op =
 | 
| -          new(I) BinaryMintOpInstr(
 | 
| -              op_kind, new(I) Value(left), new(I) Value(right),
 | 
| -              call->deopt_id());
 | 
| -      ReplaceCall(call, bin_op);
 | 
| -    }
 | 
| -  } else if (operands_type == kFloat32x4Cid) {
 | 
| -    return InlineFloat32x4BinaryOp(call, op_kind);
 | 
| -  } else if (operands_type == kInt32x4Cid) {
 | 
| -    return InlineInt32x4BinaryOp(call, op_kind);
 | 
| -  } else if (operands_type == kFloat64x2Cid) {
 | 
| -    return InlineFloat64x2BinaryOp(call, op_kind);
 | 
| -  } else if (op_kind == Token::kMOD) {
 | 
| -    ASSERT(operands_type == kSmiCid);
 | 
| -    if (right->IsConstant()) {
 | 
| -      const Object& obj = right->AsConstant()->value();
 | 
| -      if (obj.IsSmi() && Utils::IsPowerOfTwo(Smi::Cast(obj).Value())) {
 | 
| -        // Insert smi check and attach a copy of the original environment
 | 
| -        // because the smi operation can still deoptimize.
 | 
| -        InsertBefore(call,
 | 
| -                     new(I) CheckSmiInstr(new(I) Value(left),
 | 
| -                                          call->deopt_id(),
 | 
| -                                          call->token_pos()),
 | 
| -                     call->env(),
 | 
| -                     FlowGraph::kEffect);
 | 
| -        ConstantInstr* constant =
 | 
| -            flow_graph()->GetConstant(Smi::Handle(I,
 | 
| -                Smi::New(Smi::Cast(obj).Value() - 1)));
 | 
| -        BinarySmiOpInstr* bin_op =
 | 
| -            new(I) BinarySmiOpInstr(Token::kBIT_AND,
 | 
| -                                    new(I) Value(left),
 | 
| -                                    new(I) Value(constant),
 | 
| -                                    call->deopt_id());
 | 
| -        ReplaceCall(call, bin_op);
 | 
| -        return true;
 | 
| -      }
 | 
| -    }
 | 
| -    // Insert two smi checks and attach a copy of the original
 | 
| -    // environment because the smi operation can still deoptimize.
 | 
| -    AddCheckSmi(left, call->deopt_id(), call->env(), call);
 | 
| -    AddCheckSmi(right, call->deopt_id(), call->env(), call);
 | 
| -    BinarySmiOpInstr* bin_op =
 | 
| -        new(I) BinarySmiOpInstr(op_kind,
 | 
| -                                new(I) Value(left),
 | 
| -                                new(I) Value(right),
 | 
| -                                call->deopt_id());
 | 
| -    ReplaceCall(call, bin_op);
 | 
| -  } else {
 | 
| -    ASSERT(operands_type == kSmiCid);
 | 
| -    // Insert two smi checks and attach a copy of the original
 | 
| -    // environment because the smi operation can still deoptimize.
 | 
| -    AddCheckSmi(left, call->deopt_id(), call->env(), call);
 | 
| -    AddCheckSmi(right, call->deopt_id(), call->env(), call);
 | 
| -    if (left->IsConstant() &&
 | 
| -        ((op_kind == Token::kADD) || (op_kind == Token::kMUL))) {
 | 
| -      // Constant should be on the right side.
 | 
| -      Definition* temp = left;
 | 
| -      left = right;
 | 
| -      right = temp;
 | 
| -    }
 | 
| -    BinarySmiOpInstr* bin_op =
 | 
| -        new(I) BinarySmiOpInstr(
 | 
| -            op_kind,
 | 
| -            new(I) Value(left),
 | 
| -            new(I) Value(right),
 | 
| -            call->deopt_id());
 | 
| -    ReplaceCall(call, bin_op);
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryReplaceWithUnaryOp(InstanceCallInstr* call,
 | 
| -                                               Token::Kind op_kind) {
 | 
| -  ASSERT(call->ArgumentCount() == 1);
 | 
| -  Definition* input = call->ArgumentAt(0);
 | 
| -  Definition* unary_op = NULL;
 | 
| -  if (HasOnlyOneSmi(*call->ic_data())) {
 | 
| -    InsertBefore(call,
 | 
| -                 new(I) CheckSmiInstr(new(I) Value(input),
 | 
| -                                      call->deopt_id(),
 | 
| -                                      call->token_pos()),
 | 
| -                 call->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -    unary_op = new(I) UnarySmiOpInstr(
 | 
| -        op_kind, new(I) Value(input), call->deopt_id());
 | 
| -  } else if ((op_kind == Token::kBIT_NOT) &&
 | 
| -             HasOnlySmiOrMint(*call->ic_data()) &&
 | 
| -             FlowGraphCompiler::SupportsUnboxedMints()) {
 | 
| -    unary_op = new(I) UnaryMintOpInstr(
 | 
| -        op_kind, new(I) Value(input), call->deopt_id());
 | 
| -  } else if (HasOnlyOneDouble(*call->ic_data()) &&
 | 
| -             (op_kind == Token::kNEGATE) &&
 | 
| -             CanUnboxDouble()) {
 | 
| -    AddReceiverCheck(call);
 | 
| -    unary_op = new(I) UnaryDoubleOpInstr(
 | 
| -        Token::kNEGATE, new(I) Value(input), call->deopt_id());
 | 
| -  } else {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(unary_op != NULL);
 | 
| -  ReplaceCall(call, unary_op);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Using field class
 | 
| -static RawField* GetField(intptr_t class_id, const String& field_name) {
 | 
| -  Isolate* isolate = Isolate::Current();
 | 
| -  Class& cls = Class::Handle(isolate, isolate->class_table()->At(class_id));
 | 
| -  Field& field = Field::Handle(isolate);
 | 
| -  while (!cls.IsNull()) {
 | 
| -    field = cls.LookupInstanceField(field_name);
 | 
| -    if (!field.IsNull()) {
 | 
| -      return field.raw();
 | 
| -    }
 | 
| -    cls = cls.SuperClass();
 | 
| -  }
 | 
| -  return Field::null();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Use CHA to determine if the call needs a class check: if the callee's
 | 
| -// receiver is the same as the caller's receiver and there are no overriden
 | 
| -// callee functions, then no class check is needed.
 | 
| -bool FlowGraphOptimizer::InstanceCallNeedsClassCheck(
 | 
| -    InstanceCallInstr* call, RawFunction::Kind kind) const {
 | 
| -  if (!FLAG_use_cha) return true;
 | 
| -  Definition* callee_receiver = call->ArgumentAt(0);
 | 
| -  ASSERT(callee_receiver != NULL);
 | 
| -  const Function& function = flow_graph_->parsed_function()->function();
 | 
| -  if (function.IsDynamicFunction() &&
 | 
| -      callee_receiver->IsParameter() &&
 | 
| -      (callee_receiver->AsParameter()->index() == 0)) {
 | 
| -    const String& name = (kind == RawFunction::kMethodExtractor)
 | 
| -        ? String::Handle(I, Field::NameFromGetter(call->function_name()))
 | 
| -        : call->function_name();
 | 
| -    return isolate()->cha()->HasOverride(Class::Handle(I, function.Owner()),
 | 
| -                                         name);
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::InlineImplicitInstanceGetter(InstanceCallInstr* call) {
 | 
| -  ASSERT(call->HasICData());
 | 
| -  const ICData& ic_data = *call->ic_data();
 | 
| -  ASSERT(ic_data.HasOneTarget());
 | 
| -  Function& target = Function::Handle(I);
 | 
| -  GrowableArray<intptr_t> class_ids;
 | 
| -  ic_data.GetCheckAt(0, &class_ids, &target);
 | 
| -  ASSERT(class_ids.length() == 1);
 | 
| -  // Inline implicit instance getter.
 | 
| -  const String& field_name =
 | 
| -      String::Handle(I, Field::NameFromGetter(call->function_name()));
 | 
| -  const Field& field =
 | 
| -      Field::ZoneHandle(I, GetField(class_ids[0], field_name));
 | 
| -  ASSERT(!field.IsNull());
 | 
| -
 | 
| -  if (InstanceCallNeedsClassCheck(call, RawFunction::kImplicitGetter)) {
 | 
| -    AddReceiverCheck(call);
 | 
| -  }
 | 
| -  LoadFieldInstr* load = new(I) LoadFieldInstr(
 | 
| -      new(I) Value(call->ArgumentAt(0)),
 | 
| -      &field,
 | 
| -      AbstractType::ZoneHandle(I, field.type()),
 | 
| -      call->token_pos());
 | 
| -  load->set_is_immutable(field.is_final());
 | 
| -  if (field.guarded_cid() != kIllegalCid) {
 | 
| -    if (!field.is_nullable() || (field.guarded_cid() == kNullCid)) {
 | 
| -      load->set_result_cid(field.guarded_cid());
 | 
| -    }
 | 
| -    FlowGraph::AddToGuardedFields(flow_graph_->guarded_fields(), &field);
 | 
| -  }
 | 
| -
 | 
| -  // Discard the environment from the original instruction because the load
 | 
| -  // can't deoptimize.
 | 
| -  call->RemoveEnvironment();
 | 
| -  ReplaceCall(call, load);
 | 
| -
 | 
| -  if (load->result_cid() != kDynamicCid) {
 | 
| -    // Reset value types if guarded_cid was used.
 | 
| -    for (Value::Iterator it(load->input_use_list());
 | 
| -         !it.Done();
 | 
| -         it.Advance()) {
 | 
| -      it.Current()->SetReachingType(NULL);
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineFloat32x4Getter(InstanceCallInstr* call,
 | 
| -                                               MethodRecognizer::Kind getter) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  AddCheckClass(call->ArgumentAt(0),
 | 
| -                ICData::ZoneHandle(
 | 
| -                    I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  intptr_t mask = 0;
 | 
| -  if ((getter == MethodRecognizer::kFloat32x4Shuffle) ||
 | 
| -      (getter == MethodRecognizer::kFloat32x4ShuffleMix)) {
 | 
| -    // Extract shuffle mask.
 | 
| -    Definition* mask_definition = NULL;
 | 
| -    if (getter == MethodRecognizer::kFloat32x4Shuffle) {
 | 
| -      ASSERT(call->ArgumentCount() == 2);
 | 
| -      mask_definition = call->ArgumentAt(1);
 | 
| -    } else {
 | 
| -      ASSERT(getter == MethodRecognizer::kFloat32x4ShuffleMix);
 | 
| -      ASSERT(call->ArgumentCount() == 3);
 | 
| -      mask_definition = call->ArgumentAt(2);
 | 
| -    }
 | 
| -    if (!mask_definition->IsConstant()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -    ASSERT(mask_definition->IsConstant());
 | 
| -    ConstantInstr* constant_instruction = mask_definition->AsConstant();
 | 
| -    const Object& constant_mask = constant_instruction->value();
 | 
| -    if (!constant_mask.IsSmi()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -    ASSERT(constant_mask.IsSmi());
 | 
| -    mask = Smi::Cast(constant_mask).Value();
 | 
| -    if ((mask < 0) || (mask > 255)) {
 | 
| -      // Not a valid mask.
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -  if (getter == MethodRecognizer::kFloat32x4GetSignMask) {
 | 
| -    Simd32x4GetSignMaskInstr* instr = new(I) Simd32x4GetSignMaskInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  } else if (getter == MethodRecognizer::kFloat32x4ShuffleMix) {
 | 
| -    Simd32x4ShuffleMixInstr* instr = new(I) Simd32x4ShuffleMixInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        new(I) Value(call->ArgumentAt(1)),
 | 
| -        mask,
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  } else {
 | 
| -    ASSERT((getter == MethodRecognizer::kFloat32x4Shuffle)  ||
 | 
| -           (getter == MethodRecognizer::kFloat32x4ShuffleX) ||
 | 
| -           (getter == MethodRecognizer::kFloat32x4ShuffleY) ||
 | 
| -           (getter == MethodRecognizer::kFloat32x4ShuffleZ) ||
 | 
| -           (getter == MethodRecognizer::kFloat32x4ShuffleW));
 | 
| -    Simd32x4ShuffleInstr* instr = new(I) Simd32x4ShuffleInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        mask,
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  }
 | 
| -  UNREACHABLE();
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineFloat64x2Getter(InstanceCallInstr* call,
 | 
| -                                               MethodRecognizer::Kind getter) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  AddCheckClass(call->ArgumentAt(0),
 | 
| -                ICData::ZoneHandle(
 | 
| -                    I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  if ((getter == MethodRecognizer::kFloat64x2GetX) ||
 | 
| -      (getter == MethodRecognizer::kFloat64x2GetY)) {
 | 
| -    Simd64x2ShuffleInstr* instr = new(I) Simd64x2ShuffleInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        0,
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  }
 | 
| -  UNREACHABLE();
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineInt32x4Getter(InstanceCallInstr* call,
 | 
| -                                              MethodRecognizer::Kind getter) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  AddCheckClass(call->ArgumentAt(0),
 | 
| -                ICData::ZoneHandle(
 | 
| -                    I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  intptr_t mask = 0;
 | 
| -  if ((getter == MethodRecognizer::kInt32x4Shuffle) ||
 | 
| -      (getter == MethodRecognizer::kInt32x4ShuffleMix)) {
 | 
| -    // Extract shuffle mask.
 | 
| -    Definition* mask_definition = NULL;
 | 
| -    if (getter == MethodRecognizer::kInt32x4Shuffle) {
 | 
| -      ASSERT(call->ArgumentCount() == 2);
 | 
| -      mask_definition = call->ArgumentAt(1);
 | 
| -    } else {
 | 
| -      ASSERT(getter == MethodRecognizer::kInt32x4ShuffleMix);
 | 
| -      ASSERT(call->ArgumentCount() == 3);
 | 
| -      mask_definition = call->ArgumentAt(2);
 | 
| -    }
 | 
| -    if (!mask_definition->IsConstant()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -    ASSERT(mask_definition->IsConstant());
 | 
| -    ConstantInstr* constant_instruction = mask_definition->AsConstant();
 | 
| -    const Object& constant_mask = constant_instruction->value();
 | 
| -    if (!constant_mask.IsSmi()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -    ASSERT(constant_mask.IsSmi());
 | 
| -    mask = Smi::Cast(constant_mask).Value();
 | 
| -    if ((mask < 0) || (mask > 255)) {
 | 
| -      // Not a valid mask.
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -  if (getter == MethodRecognizer::kInt32x4GetSignMask) {
 | 
| -    Simd32x4GetSignMaskInstr* instr = new(I) Simd32x4GetSignMaskInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  } else if (getter == MethodRecognizer::kInt32x4ShuffleMix) {
 | 
| -    Simd32x4ShuffleMixInstr* instr = new(I) Simd32x4ShuffleMixInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        new(I) Value(call->ArgumentAt(1)),
 | 
| -        mask,
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  } else if (getter == MethodRecognizer::kInt32x4Shuffle) {
 | 
| -    Simd32x4ShuffleInstr* instr = new(I) Simd32x4ShuffleInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        mask,
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  } else {
 | 
| -    Int32x4GetFlagInstr* instr = new(I) Int32x4GetFlagInstr(
 | 
| -        getter,
 | 
| -        new(I) Value(call->ArgumentAt(0)),
 | 
| -        call->deopt_id());
 | 
| -    ReplaceCall(call, instr);
 | 
| -    return true;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineFloat32x4BinaryOp(InstanceCallInstr* call,
 | 
| -                                                 Token::Kind op_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(call->ArgumentCount() == 2);
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -  // Type check left.
 | 
| -  AddCheckClass(left,
 | 
| -                ICData::ZoneHandle(
 | 
| -                    I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  // Type check right.
 | 
| -  AddCheckClass(right,
 | 
| -                ICData::ZoneHandle(
 | 
| -                    I, call->ic_data()->AsUnaryClassChecksForArgNr(1)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  // Replace call.
 | 
| -  BinaryFloat32x4OpInstr* float32x4_bin_op =
 | 
| -      new(I) BinaryFloat32x4OpInstr(
 | 
| -          op_kind, new(I) Value(left), new(I) Value(right),
 | 
| -          call->deopt_id());
 | 
| -  ReplaceCall(call, float32x4_bin_op);
 | 
| -
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineInt32x4BinaryOp(InstanceCallInstr* call,
 | 
| -                                                Token::Kind op_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(call->ArgumentCount() == 2);
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -  // Type check left.
 | 
| -  AddCheckClass(left,
 | 
| -                ICData::ZoneHandle(
 | 
| -                    I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  // Type check right.
 | 
| -  AddCheckClass(right,
 | 
| -                ICData::ZoneHandle(I,
 | 
| -                    call->ic_data()->AsUnaryClassChecksForArgNr(1)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  // Replace call.
 | 
| -  BinaryInt32x4OpInstr* int32x4_bin_op =
 | 
| -      new(I) BinaryInt32x4OpInstr(
 | 
| -          op_kind, new(I) Value(left), new(I) Value(right),
 | 
| -          call->deopt_id());
 | 
| -  ReplaceCall(call, int32x4_bin_op);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineFloat64x2BinaryOp(InstanceCallInstr* call,
 | 
| -                                                 Token::Kind op_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(call->ArgumentCount() == 2);
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -  // Type check left.
 | 
| -  AddCheckClass(left,
 | 
| -                ICData::ZoneHandle(
 | 
| -                    call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  // Type check right.
 | 
| -  AddCheckClass(right,
 | 
| -                ICData::ZoneHandle(
 | 
| -                    call->ic_data()->AsUnaryClassChecksForArgNr(1)),
 | 
| -                call->deopt_id(),
 | 
| -                call->env(),
 | 
| -                call);
 | 
| -  // Replace call.
 | 
| -  BinaryFloat64x2OpInstr* float64x2_bin_op =
 | 
| -      new(I) BinaryFloat64x2OpInstr(
 | 
| -          op_kind, new(I) Value(left), new(I) Value(right),
 | 
| -          call->deopt_id());
 | 
| -  ReplaceCall(call, float64x2_bin_op);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Only unique implicit instance getters can be currently handled.
 | 
| -bool FlowGraphOptimizer::TryInlineInstanceGetter(InstanceCallInstr* call) {
 | 
| -  ASSERT(call->HasICData());
 | 
| -  const ICData& ic_data = *call->ic_data();
 | 
| -  if (ic_data.NumberOfUsedChecks() == 0) {
 | 
| -    // No type feedback collected.
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  if (!ic_data.HasOneTarget()) {
 | 
| -    // Polymorphic sites are inlined like normal methods by conventional
 | 
| -    // inlining in FlowGraphInliner.
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  const Function& target = Function::Handle(I, ic_data.GetTargetAt(0));
 | 
| -  if (target.kind() != RawFunction::kImplicitGetter) {
 | 
| -    // Non-implicit getters are inlined like normal methods by conventional
 | 
| -    // inlining in FlowGraphInliner.
 | 
| -    return false;
 | 
| -  }
 | 
| -  InlineImplicitInstanceGetter(call);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryReplaceInstanceCallWithInline(
 | 
| -    InstanceCallInstr* call) {
 | 
| -  ASSERT(call->HasICData());
 | 
| -  Function& target = Function::Handle(I);
 | 
| -  GrowableArray<intptr_t> class_ids;
 | 
| -  call->ic_data()->GetCheckAt(0, &class_ids, &target);
 | 
| -  const intptr_t receiver_cid = class_ids[0];
 | 
| -
 | 
| -  TargetEntryInstr* entry;
 | 
| -  Definition* last;
 | 
| -  if (!TryInlineRecognizedMethod(receiver_cid,
 | 
| -                                 target,
 | 
| -                                 call,
 | 
| -                                 call->ArgumentAt(0),
 | 
| -                                 call->token_pos(),
 | 
| -                                 *call->ic_data(),
 | 
| -                                 &entry, &last)) {
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  // Insert receiver class check.
 | 
| -  AddReceiverCheck(call);
 | 
| -  // Remove the original push arguments.
 | 
| -  for (intptr_t i = 0; i < call->ArgumentCount(); ++i) {
 | 
| -    PushArgumentInstr* push = call->PushArgumentAt(i);
 | 
| -    push->ReplaceUsesWith(push->value()->definition());
 | 
| -    push->RemoveFromGraph();
 | 
| -  }
 | 
| -  // Replace all uses of this definition with the result.
 | 
| -  call->ReplaceUsesWith(last);
 | 
| -  // Finally insert the sequence other definition in place of this one in the
 | 
| -  // graph.
 | 
| -  call->previous()->LinkTo(entry->next());
 | 
| -  entry->UnuseAllInputs();  // Entry block is not in the graph.
 | 
| -  last->LinkTo(call);
 | 
| -  // Remove through the iterator.
 | 
| -  ASSERT(current_iterator()->Current() == call);
 | 
| -  current_iterator()->RemoveCurrentFromGraph();
 | 
| -  call->set_previous(NULL);
 | 
| -  call->set_next(NULL);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Returns the LoadIndexedInstr.
 | 
| -Definition* FlowGraphOptimizer::PrepareInlineStringIndexOp(
 | 
| -    Instruction* call,
 | 
| -    intptr_t cid,
 | 
| -    Definition* str,
 | 
| -    Definition* index,
 | 
| -    Instruction* cursor) {
 | 
| -
 | 
| -  cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                  new(I) CheckSmiInstr(
 | 
| -                                      new(I) Value(index),
 | 
| -                                      call->deopt_id(),
 | 
| -                                      call->token_pos()),
 | 
| -                                  call->env(),
 | 
| -                                  FlowGraph::kEffect);
 | 
| -
 | 
| -  // Load the length of the string.
 | 
| -  // Treat length loads as mutable (i.e. affected by side effects) to avoid
 | 
| -  // hoisting them since we can't hoist the preceding class-check. This
 | 
| -  // is because of externalization of strings that affects their class-id.
 | 
| -  LoadFieldInstr* length = new(I) LoadFieldInstr(
 | 
| -      new(I) Value(str),
 | 
| -      String::length_offset(),
 | 
| -      Type::ZoneHandle(I, Type::SmiType()),
 | 
| -      str->token_pos());
 | 
| -  length->set_result_cid(kSmiCid);
 | 
| -  length->set_recognized_kind(MethodRecognizer::kStringBaseLength);
 | 
| -
 | 
| -  cursor = flow_graph()->AppendTo(cursor, length, NULL, FlowGraph::kValue);
 | 
| -  // Bounds check.
 | 
| -  cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                   new(I) CheckArrayBoundInstr(
 | 
| -                                       new(I) Value(length),
 | 
| -                                       new(I) Value(index),
 | 
| -                                       call->deopt_id()),
 | 
| -                                   call->env(),
 | 
| -                                   FlowGraph::kEffect);
 | 
| -
 | 
| -  LoadIndexedInstr* load_indexed = new(I) LoadIndexedInstr(
 | 
| -      new(I) Value(str),
 | 
| -      new(I) Value(index),
 | 
| -      Instance::ElementSizeFor(cid),
 | 
| -      cid,
 | 
| -      Isolate::kNoDeoptId,
 | 
| -      call->token_pos());
 | 
| -
 | 
| -  cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                  load_indexed,
 | 
| -                                  NULL,
 | 
| -                                  FlowGraph::kValue);
 | 
| -  ASSERT(cursor == load_indexed);
 | 
| -  return load_indexed;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineStringCodeUnitAt(
 | 
| -    Instruction* call,
 | 
| -    intptr_t cid,
 | 
| -    TargetEntryInstr** entry,
 | 
| -    Definition** last) {
 | 
| -  // TODO(johnmccutchan): Handle external strings in PrepareInlineStringIndexOp.
 | 
| -  if (RawObject::IsExternalStringClassId(cid)) {
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  Definition* str = call->ArgumentAt(0);
 | 
| -  Definition* index = call->ArgumentAt(1);
 | 
| -
 | 
| -  *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(),
 | 
| -                                   call->GetBlock()->try_index());
 | 
| -  (*entry)->InheritDeoptTarget(I, call);
 | 
| -
 | 
| -  *last = PrepareInlineStringIndexOp(call, cid, str, index, *entry);
 | 
| -
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineStringBaseCharAt(
 | 
| -    Instruction* call,
 | 
| -    intptr_t cid,
 | 
| -    TargetEntryInstr** entry,
 | 
| -    Definition** last) {
 | 
| -  // TODO(johnmccutchan): Handle external strings in PrepareInlineStringIndexOp.
 | 
| -  if (RawObject::IsExternalStringClassId(cid) || cid != kOneByteStringCid) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  Definition* str = call->ArgumentAt(0);
 | 
| -  Definition* index = call->ArgumentAt(1);
 | 
| -
 | 
| -  *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(),
 | 
| -                                   call->GetBlock()->try_index());
 | 
| -  (*entry)->InheritDeoptTarget(I, call);
 | 
| -
 | 
| -  *last = PrepareInlineStringIndexOp(call, cid, str, index, *entry);
 | 
| -
 | 
| -  StringFromCharCodeInstr* char_at = new(I) StringFromCharCodeInstr(
 | 
| -      new(I) Value(*last), cid);
 | 
| -
 | 
| -  flow_graph()->AppendTo(*last, char_at, NULL, FlowGraph::kValue);
 | 
| -  *last = char_at;
 | 
| -
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineDoubleOp(
 | 
| -    Token::Kind op_kind,
 | 
| -    Instruction* call,
 | 
| -    TargetEntryInstr** entry,
 | 
| -    Definition** last) {
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* right = call->ArgumentAt(1);
 | 
| -
 | 
| -  *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(),
 | 
| -                                   call->GetBlock()->try_index());
 | 
| -  (*entry)->InheritDeoptTarget(I, call);
 | 
| -  // Arguments are checked. No need for class check.
 | 
| -  BinaryDoubleOpInstr* double_bin_op =
 | 
| -      new(I) BinaryDoubleOpInstr(op_kind,
 | 
| -                                 new(I) Value(left),
 | 
| -                                 new(I) Value(right),
 | 
| -                                 call->deopt_id(), call->token_pos());
 | 
| -  flow_graph()->AppendTo(*entry, double_bin_op, call->env(), FlowGraph::kValue);
 | 
| -  *last = double_bin_op;
 | 
| -
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::ReplaceWithMathCFunction(
 | 
| -    InstanceCallInstr* call,
 | 
| -    MethodRecognizer::Kind recognized_kind) {
 | 
| -  AddReceiverCheck(call);
 | 
| -  ZoneGrowableArray<Value*>* args =
 | 
| -      new(I) ZoneGrowableArray<Value*>(call->ArgumentCount());
 | 
| -  for (intptr_t i = 0; i < call->ArgumentCount(); i++) {
 | 
| -    args->Add(new(I) Value(call->ArgumentAt(i)));
 | 
| -  }
 | 
| -  InvokeMathCFunctionInstr* invoke =
 | 
| -      new(I) InvokeMathCFunctionInstr(args,
 | 
| -                                      call->deopt_id(),
 | 
| -                                      recognized_kind,
 | 
| -                                      call->token_pos());
 | 
| -  ReplaceCall(call, invoke);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool IsSupportedByteArrayViewCid(intptr_t cid) {
 | 
| -  switch (cid) {
 | 
| -    case kTypedDataInt8ArrayCid:
 | 
| -    case kTypedDataUint8ArrayCid:
 | 
| -    case kExternalTypedDataUint8ArrayCid:
 | 
| -    case kTypedDataUint8ClampedArrayCid:
 | 
| -    case kExternalTypedDataUint8ClampedArrayCid:
 | 
| -    case kTypedDataInt16ArrayCid:
 | 
| -    case kTypedDataUint16ArrayCid:
 | 
| -    case kTypedDataInt32ArrayCid:
 | 
| -    case kTypedDataUint32ArrayCid:
 | 
| -    case kTypedDataFloat32ArrayCid:
 | 
| -    case kTypedDataFloat64ArrayCid:
 | 
| -    case kTypedDataFloat32x4ArrayCid:
 | 
| -    case kTypedDataInt32x4ArrayCid:
 | 
| -      return true;
 | 
| -    default:
 | 
| -      return false;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Inline only simple, frequently called core library methods.
 | 
| -bool FlowGraphOptimizer::TryInlineInstanceMethod(InstanceCallInstr* call) {
 | 
| -  ASSERT(call->HasICData());
 | 
| -  const ICData& ic_data = *call->ic_data();
 | 
| -  if ((ic_data.NumberOfUsedChecks() == 0) || !ic_data.HasOneTarget()) {
 | 
| -    // No type feedback collected or multiple targets found.
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  Function& target = Function::Handle(I);
 | 
| -  GrowableArray<intptr_t> class_ids;
 | 
| -  ic_data.GetCheckAt(0, &class_ids, &target);
 | 
| -  MethodRecognizer::Kind recognized_kind =
 | 
| -      MethodRecognizer::RecognizeKind(target);
 | 
| -
 | 
| -  if ((recognized_kind == MethodRecognizer::kGrowableArraySetData) &&
 | 
| -      (ic_data.NumberOfChecks() == 1) &&
 | 
| -      (class_ids[0] == kGrowableObjectArrayCid)) {
 | 
| -    // This is an internal method, no need to check argument types.
 | 
| -    Definition* array = call->ArgumentAt(0);
 | 
| -    Definition* value = call->ArgumentAt(1);
 | 
| -    StoreInstanceFieldInstr* store = new(I) StoreInstanceFieldInstr(
 | 
| -        GrowableObjectArray::data_offset(),
 | 
| -        new(I) Value(array),
 | 
| -        new(I) Value(value),
 | 
| -        kEmitStoreBarrier,
 | 
| -        call->token_pos());
 | 
| -    ReplaceCall(call, store);
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  if ((recognized_kind == MethodRecognizer::kGrowableArraySetLength) &&
 | 
| -      (ic_data.NumberOfChecks() == 1) &&
 | 
| -      (class_ids[0] == kGrowableObjectArrayCid)) {
 | 
| -    // This is an internal method, no need to check argument types nor
 | 
| -    // range.
 | 
| -    Definition* array = call->ArgumentAt(0);
 | 
| -    Definition* value = call->ArgumentAt(1);
 | 
| -    StoreInstanceFieldInstr* store = new(I) StoreInstanceFieldInstr(
 | 
| -        GrowableObjectArray::length_offset(),
 | 
| -        new(I) Value(array),
 | 
| -        new(I) Value(value),
 | 
| -        kNoStoreBarrier,
 | 
| -        call->token_pos());
 | 
| -    ReplaceCall(call, store);
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  if (((recognized_kind == MethodRecognizer::kStringBaseCodeUnitAt) ||
 | 
| -       (recognized_kind == MethodRecognizer::kStringBaseCharAt)) &&
 | 
| -      (ic_data.NumberOfChecks() == 1) &&
 | 
| -      ((class_ids[0] == kOneByteStringCid) ||
 | 
| -       (class_ids[0] == kTwoByteStringCid))) {
 | 
| -    return TryReplaceInstanceCallWithInline(call);
 | 
| -  }
 | 
| -
 | 
| -  if ((class_ids[0] == kOneByteStringCid) && (ic_data.NumberOfChecks() == 1)) {
 | 
| -    if (recognized_kind == MethodRecognizer::kOneByteStringSetAt) {
 | 
| -      // This is an internal method, no need to check argument types nor
 | 
| -      // range.
 | 
| -      Definition* str = call->ArgumentAt(0);
 | 
| -      Definition* index = call->ArgumentAt(1);
 | 
| -      Definition* value = call->ArgumentAt(2);
 | 
| -      StoreIndexedInstr* store_op = new(I) StoreIndexedInstr(
 | 
| -          new(I) Value(str),
 | 
| -          new(I) Value(index),
 | 
| -          new(I) Value(value),
 | 
| -          kNoStoreBarrier,
 | 
| -          1,  // Index scale
 | 
| -          kOneByteStringCid,
 | 
| -          call->deopt_id(),
 | 
| -          call->token_pos());
 | 
| -      ReplaceCall(call, store_op);
 | 
| -      return true;
 | 
| -    }
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  if (CanUnboxDouble() &&
 | 
| -      (recognized_kind == MethodRecognizer::kIntegerToDouble) &&
 | 
| -      (ic_data.NumberOfChecks() == 1)) {
 | 
| -    if (class_ids[0] == kSmiCid) {
 | 
| -      AddReceiverCheck(call);
 | 
| -      ReplaceCall(call,
 | 
| -                  new(I) SmiToDoubleInstr(
 | 
| -                      new(I) Value(call->ArgumentAt(0)),
 | 
| -                      call->token_pos()));
 | 
| -      return true;
 | 
| -    } else if ((class_ids[0] == kMintCid) && CanConvertUnboxedMintToDouble()) {
 | 
| -      AddReceiverCheck(call);
 | 
| -      ReplaceCall(call,
 | 
| -                  new(I) MintToDoubleInstr(new(I) Value(call->ArgumentAt(0)),
 | 
| -                                           call->deopt_id()));
 | 
| -      return true;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (class_ids[0] == kDoubleCid) {
 | 
| -    if (!CanUnboxDouble()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -    switch (recognized_kind) {
 | 
| -      case MethodRecognizer::kDoubleToInteger: {
 | 
| -        AddReceiverCheck(call);
 | 
| -        ASSERT(call->HasICData());
 | 
| -        const ICData& ic_data = *call->ic_data();
 | 
| -        Definition* input = call->ArgumentAt(0);
 | 
| -        Definition* d2i_instr = NULL;
 | 
| -        if (ic_data.HasDeoptReason(ICData::kDeoptDoubleToSmi)) {
 | 
| -          // Do not repeatedly deoptimize because result didn't fit into Smi.
 | 
| -          d2i_instr =  new(I) DoubleToIntegerInstr(
 | 
| -              new(I) Value(input), call);
 | 
| -        } else {
 | 
| -          // Optimistically assume result fits into Smi.
 | 
| -          d2i_instr = new(I) DoubleToSmiInstr(
 | 
| -              new(I) Value(input), call->deopt_id());
 | 
| -        }
 | 
| -        ReplaceCall(call, d2i_instr);
 | 
| -        return true;
 | 
| -      }
 | 
| -      case MethodRecognizer::kDoubleMod:
 | 
| -      case MethodRecognizer::kDoubleRound:
 | 
| -        ReplaceWithMathCFunction(call, recognized_kind);
 | 
| -        return true;
 | 
| -      case MethodRecognizer::kDoubleTruncate:
 | 
| -      case MethodRecognizer::kDoubleFloor:
 | 
| -      case MethodRecognizer::kDoubleCeil:
 | 
| -        if (!TargetCPUFeatures::double_truncate_round_supported()) {
 | 
| -          ReplaceWithMathCFunction(call, recognized_kind);
 | 
| -        } else {
 | 
| -          AddReceiverCheck(call);
 | 
| -          DoubleToDoubleInstr* d2d_instr =
 | 
| -              new(I) DoubleToDoubleInstr(new(I) Value(call->ArgumentAt(0)),
 | 
| -                                         recognized_kind, call->deopt_id());
 | 
| -          ReplaceCall(call, d2d_instr);
 | 
| -        }
 | 
| -        return true;
 | 
| -      case MethodRecognizer::kDoubleAdd:
 | 
| -      case MethodRecognizer::kDoubleSub:
 | 
| -      case MethodRecognizer::kDoubleMul:
 | 
| -      case MethodRecognizer::kDoubleDiv:
 | 
| -        return TryReplaceInstanceCallWithInline(call);
 | 
| -      default:
 | 
| -        // Unsupported method.
 | 
| -        return false;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (IsSupportedByteArrayViewCid(class_ids[0]) &&
 | 
| -      (ic_data.NumberOfChecks() == 1)) {
 | 
| -    // For elements that may not fit into a smi on all platforms, check if
 | 
| -    // elements fit into a smi or the platform supports unboxed mints.
 | 
| -    if ((recognized_kind == MethodRecognizer::kByteArrayBaseGetInt32) ||
 | 
| -        (recognized_kind == MethodRecognizer::kByteArrayBaseGetUint32) ||
 | 
| -        (recognized_kind == MethodRecognizer::kByteArrayBaseSetInt32) ||
 | 
| -        (recognized_kind == MethodRecognizer::kByteArrayBaseSetUint32)) {
 | 
| -      if (!CanUnboxInt32()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    if ((recognized_kind == MethodRecognizer::kByteArrayBaseGetFloat32) ||
 | 
| -        (recognized_kind == MethodRecognizer::kByteArrayBaseGetFloat64) ||
 | 
| -        (recognized_kind == MethodRecognizer::kByteArrayBaseSetFloat32) ||
 | 
| -        (recognized_kind == MethodRecognizer::kByteArrayBaseSetFloat64)) {
 | 
| -      if (!CanUnboxDouble()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    switch (recognized_kind) {
 | 
| -      // ByteArray getters.
 | 
| -      case MethodRecognizer::kByteArrayBaseGetInt8:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataInt8ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetUint8:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataUint8ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetInt16:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataInt16ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetUint16:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataUint16ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetInt32:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataInt32ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetUint32:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataUint32ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetFloat32:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataFloat32ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetFloat64:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataFloat64ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetFloat32x4:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataFloat32x4ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseGetInt32x4:
 | 
| -        return BuildByteArrayViewLoad(call, kTypedDataInt32x4ArrayCid);
 | 
| -
 | 
| -      // ByteArray setters.
 | 
| -      case MethodRecognizer::kByteArrayBaseSetInt8:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataInt8ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetUint8:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataUint8ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetInt16:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataInt16ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetUint16:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataUint16ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetInt32:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataInt32ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetUint32:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataUint32ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetFloat32:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataFloat32ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetFloat64:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataFloat64ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetFloat32x4:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataFloat32x4ArrayCid);
 | 
| -      case MethodRecognizer::kByteArrayBaseSetInt32x4:
 | 
| -        return BuildByteArrayViewStore(call, kTypedDataInt32x4ArrayCid);
 | 
| -      default:
 | 
| -        // Unsupported method.
 | 
| -        return false;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if ((class_ids[0] == kFloat32x4Cid) && (ic_data.NumberOfChecks() == 1)) {
 | 
| -    return TryInlineFloat32x4Method(call, recognized_kind);
 | 
| -  }
 | 
| -
 | 
| -  if ((class_ids[0] == kInt32x4Cid) && (ic_data.NumberOfChecks() == 1)) {
 | 
| -    return TryInlineInt32x4Method(call, recognized_kind);
 | 
| -  }
 | 
| -
 | 
| -  if ((class_ids[0] == kFloat64x2Cid) && (ic_data.NumberOfChecks() == 1)) {
 | 
| -    return TryInlineFloat64x2Method(call, recognized_kind);
 | 
| -  }
 | 
| -
 | 
| -  if (recognized_kind == MethodRecognizer::kIntegerLeftShiftWithMask32) {
 | 
| -    ASSERT(call->ArgumentCount() == 3);
 | 
| -    ASSERT(ic_data.NumArgsTested() == 2);
 | 
| -    Definition* value = call->ArgumentAt(0);
 | 
| -    Definition* count = call->ArgumentAt(1);
 | 
| -    Definition* int32_mask = call->ArgumentAt(2);
 | 
| -    if (HasOnlyTwoOf(ic_data, kSmiCid)) {
 | 
| -      if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      // We cannot overflow. The input value must be a Smi
 | 
| -      AddCheckSmi(value, call->deopt_id(), call->env(), call);
 | 
| -      AddCheckSmi(count, call->deopt_id(), call->env(), call);
 | 
| -      ASSERT(int32_mask->IsConstant());
 | 
| -      const Integer& mask_literal = Integer::Cast(
 | 
| -          int32_mask->AsConstant()->value());
 | 
| -      const int64_t mask_value = mask_literal.AsInt64Value();
 | 
| -      ASSERT(mask_value >= 0);
 | 
| -      if (mask_value > Smi::kMaxValue) {
 | 
| -        // The result will not be Smi.
 | 
| -        return false;
 | 
| -      }
 | 
| -      BinarySmiOpInstr* left_shift =
 | 
| -          new(I) BinarySmiOpInstr(Token::kSHL,
 | 
| -                                  new(I) Value(value),
 | 
| -                                  new(I) Value(count),
 | 
| -                                  call->deopt_id());
 | 
| -      left_shift->mark_truncating();
 | 
| -      if ((kBitsPerWord == 32) && (mask_value == 0xffffffffLL)) {
 | 
| -        // No BIT_AND operation needed.
 | 
| -        ReplaceCall(call, left_shift);
 | 
| -      } else {
 | 
| -        InsertBefore(call, left_shift, call->env(), FlowGraph::kValue);
 | 
| -        BinarySmiOpInstr* bit_and =
 | 
| -            new(I) BinarySmiOpInstr(Token::kBIT_AND,
 | 
| -                                    new(I) Value(left_shift),
 | 
| -                                    new(I) Value(int32_mask),
 | 
| -                                    call->deopt_id());
 | 
| -        ReplaceCall(call, bit_and);
 | 
| -      }
 | 
| -      return true;
 | 
| -    }
 | 
| -
 | 
| -    if (HasTwoMintOrSmi(ic_data) &&
 | 
| -        HasOnlyOneSmi(ICData::Handle(I,
 | 
| -                                     ic_data.AsUnaryClassChecksForArgNr(1)))) {
 | 
| -      if (!FlowGraphCompiler::SupportsUnboxedMints() ||
 | 
| -          ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) {
 | 
| -        return false;
 | 
| -      }
 | 
| -      ShiftMintOpInstr* left_shift =
 | 
| -          new(I) ShiftMintOpInstr(Token::kSHL,
 | 
| -                                  new(I) Value(value),
 | 
| -                                  new(I) Value(count),
 | 
| -                                  call->deopt_id());
 | 
| -      InsertBefore(call, left_shift, call->env(), FlowGraph::kValue);
 | 
| -      BinaryMintOpInstr* bit_and =
 | 
| -          new(I) BinaryMintOpInstr(Token::kBIT_AND,
 | 
| -                                   new(I) Value(left_shift),
 | 
| -                                   new(I) Value(int32_mask),
 | 
| -                                   call->deopt_id());
 | 
| -      ReplaceCall(call, bit_and);
 | 
| -      return true;
 | 
| -    }
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineFloat32x4Constructor(
 | 
| -    StaticCallInstr* call,
 | 
| -    MethodRecognizer::Kind recognized_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  if (recognized_kind == MethodRecognizer::kFloat32x4Zero) {
 | 
| -    Float32x4ZeroInstr* zero = new(I) Float32x4ZeroInstr();
 | 
| -    ReplaceCall(call, zero);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kFloat32x4Splat) {
 | 
| -    Float32x4SplatInstr* splat =
 | 
| -        new(I) Float32x4SplatInstr(
 | 
| -            new(I) Value(call->ArgumentAt(1)), call->deopt_id());
 | 
| -    ReplaceCall(call, splat);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kFloat32x4Constructor) {
 | 
| -    Float32x4ConstructorInstr* con =
 | 
| -        new(I) Float32x4ConstructorInstr(
 | 
| -            new(I) Value(call->ArgumentAt(1)),
 | 
| -            new(I) Value(call->ArgumentAt(2)),
 | 
| -            new(I) Value(call->ArgumentAt(3)),
 | 
| -            new(I) Value(call->ArgumentAt(4)),
 | 
| -            call->deopt_id());
 | 
| -    ReplaceCall(call, con);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kFloat32x4FromInt32x4Bits) {
 | 
| -    Int32x4ToFloat32x4Instr* cast =
 | 
| -        new(I) Int32x4ToFloat32x4Instr(
 | 
| -            new(I) Value(call->ArgumentAt(1)), call->deopt_id());
 | 
| -    ReplaceCall(call, cast);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kFloat32x4FromFloat64x2) {
 | 
| -    Float64x2ToFloat32x4Instr* cast =
 | 
| -        new(I) Float64x2ToFloat32x4Instr(
 | 
| -            new(I) Value(call->ArgumentAt(1)), call->deopt_id());
 | 
| -    ReplaceCall(call, cast);
 | 
| -    return true;
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineFloat64x2Constructor(
 | 
| -    StaticCallInstr* call,
 | 
| -    MethodRecognizer::Kind recognized_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  if (recognized_kind == MethodRecognizer::kFloat64x2Zero) {
 | 
| -    Float64x2ZeroInstr* zero = new(I) Float64x2ZeroInstr();
 | 
| -    ReplaceCall(call, zero);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kFloat64x2Splat) {
 | 
| -    Float64x2SplatInstr* splat =
 | 
| -        new(I) Float64x2SplatInstr(
 | 
| -            new(I) Value(call->ArgumentAt(1)), call->deopt_id());
 | 
| -    ReplaceCall(call, splat);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kFloat64x2Constructor) {
 | 
| -    Float64x2ConstructorInstr* con =
 | 
| -        new(I) Float64x2ConstructorInstr(
 | 
| -            new(I) Value(call->ArgumentAt(1)),
 | 
| -            new(I) Value(call->ArgumentAt(2)),
 | 
| -            call->deopt_id());
 | 
| -    ReplaceCall(call, con);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kFloat64x2FromFloat32x4) {
 | 
| -    Float32x4ToFloat64x2Instr* cast =
 | 
| -        new(I) Float32x4ToFloat64x2Instr(
 | 
| -            new(I) Value(call->ArgumentAt(1)), call->deopt_id());
 | 
| -    ReplaceCall(call, cast);
 | 
| -    return true;
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineInt32x4Constructor(
 | 
| -    StaticCallInstr* call,
 | 
| -    MethodRecognizer::Kind recognized_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  if (recognized_kind == MethodRecognizer::kInt32x4BoolConstructor) {
 | 
| -    Int32x4BoolConstructorInstr* con =
 | 
| -        new(I) Int32x4BoolConstructorInstr(
 | 
| -            new(I) Value(call->ArgumentAt(1)),
 | 
| -            new(I) Value(call->ArgumentAt(2)),
 | 
| -            new(I) Value(call->ArgumentAt(3)),
 | 
| -            new(I) Value(call->ArgumentAt(4)),
 | 
| -            call->deopt_id());
 | 
| -    ReplaceCall(call, con);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kInt32x4FromFloat32x4Bits) {
 | 
| -    Float32x4ToInt32x4Instr* cast =
 | 
| -        new(I) Float32x4ToInt32x4Instr(
 | 
| -            new(I) Value(call->ArgumentAt(1)), call->deopt_id());
 | 
| -    ReplaceCall(call, cast);
 | 
| -    return true;
 | 
| -  } else if (recognized_kind == MethodRecognizer::kInt32x4Constructor) {
 | 
| -    Int32x4ConstructorInstr* con =
 | 
| -        new(I) Int32x4ConstructorInstr(
 | 
| -            new(I) Value(call->ArgumentAt(1)),
 | 
| -            new(I) Value(call->ArgumentAt(2)),
 | 
| -            new(I) Value(call->ArgumentAt(3)),
 | 
| -            new(I) Value(call->ArgumentAt(4)),
 | 
| -            call->deopt_id());
 | 
| -    ReplaceCall(call, con);
 | 
| -    return true;
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineFloat32x4Method(
 | 
| -    InstanceCallInstr* call,
 | 
| -    MethodRecognizer::Kind recognized_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(call->HasICData());
 | 
| -  switch (recognized_kind) {
 | 
| -    case MethodRecognizer::kFloat32x4ShuffleX:
 | 
| -    case MethodRecognizer::kFloat32x4ShuffleY:
 | 
| -    case MethodRecognizer::kFloat32x4ShuffleZ:
 | 
| -    case MethodRecognizer::kFloat32x4ShuffleW:
 | 
| -    case MethodRecognizer::kFloat32x4GetSignMask:
 | 
| -      ASSERT(call->ic_data()->HasReceiverClassId(kFloat32x4Cid));
 | 
| -      ASSERT(call->ic_data()->HasOneTarget());
 | 
| -      return InlineFloat32x4Getter(call, recognized_kind);
 | 
| -
 | 
| -    case MethodRecognizer::kFloat32x4Equal:
 | 
| -    case MethodRecognizer::kFloat32x4GreaterThan:
 | 
| -    case MethodRecognizer::kFloat32x4GreaterThanOrEqual:
 | 
| -    case MethodRecognizer::kFloat32x4LessThan:
 | 
| -    case MethodRecognizer::kFloat32x4LessThanOrEqual:
 | 
| -    case MethodRecognizer::kFloat32x4NotEqual: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      Definition* right = call->ArgumentAt(1);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      // Replace call.
 | 
| -      Float32x4ComparisonInstr* cmp =
 | 
| -          new(I) Float32x4ComparisonInstr(recognized_kind,
 | 
| -                                          new(I) Value(left),
 | 
| -                                          new(I) Value(right),
 | 
| -                                          call->deopt_id());
 | 
| -      ReplaceCall(call, cmp);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat32x4Min:
 | 
| -    case MethodRecognizer::kFloat32x4Max: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      Definition* right = call->ArgumentAt(1);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Float32x4MinMaxInstr* minmax =
 | 
| -          new(I) Float32x4MinMaxInstr(
 | 
| -              recognized_kind,
 | 
| -              new(I) Value(left),
 | 
| -              new(I) Value(right),
 | 
| -              call->deopt_id());
 | 
| -      ReplaceCall(call, minmax);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat32x4Scale: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      Definition* right = call->ArgumentAt(1);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      // Left and right values are swapped when handed to the instruction,
 | 
| -      // this is done so that the double value is loaded into the output
 | 
| -      // register and can be destroyed.
 | 
| -      Float32x4ScaleInstr* scale =
 | 
| -          new(I) Float32x4ScaleInstr(recognized_kind,
 | 
| -                                     new(I) Value(right),
 | 
| -                                     new(I) Value(left),
 | 
| -                                     call->deopt_id());
 | 
| -      ReplaceCall(call, scale);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat32x4Sqrt:
 | 
| -    case MethodRecognizer::kFloat32x4ReciprocalSqrt:
 | 
| -    case MethodRecognizer::kFloat32x4Reciprocal: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Float32x4SqrtInstr* sqrt =
 | 
| -          new(I) Float32x4SqrtInstr(recognized_kind,
 | 
| -                                    new(I) Value(left),
 | 
| -                                    call->deopt_id());
 | 
| -      ReplaceCall(call, sqrt);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat32x4WithX:
 | 
| -    case MethodRecognizer::kFloat32x4WithY:
 | 
| -    case MethodRecognizer::kFloat32x4WithZ:
 | 
| -    case MethodRecognizer::kFloat32x4WithW: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      Definition* right = call->ArgumentAt(1);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Float32x4WithInstr* with = new(I) Float32x4WithInstr(recognized_kind,
 | 
| -                                                           new(I) Value(left),
 | 
| -                                                           new(I) Value(right),
 | 
| -                                                           call->deopt_id());
 | 
| -      ReplaceCall(call, with);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat32x4Absolute:
 | 
| -    case MethodRecognizer::kFloat32x4Negate: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Float32x4ZeroArgInstr* zeroArg =
 | 
| -          new(I) Float32x4ZeroArgInstr(
 | 
| -              recognized_kind, new(I) Value(left), call->deopt_id());
 | 
| -      ReplaceCall(call, zeroArg);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat32x4Clamp: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      Definition* lower = call->ArgumentAt(1);
 | 
| -      Definition* upper = call->ArgumentAt(2);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Float32x4ClampInstr* clamp = new(I) Float32x4ClampInstr(
 | 
| -          new(I) Value(left),
 | 
| -          new(I) Value(lower),
 | 
| -          new(I) Value(upper),
 | 
| -          call->deopt_id());
 | 
| -      ReplaceCall(call, clamp);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat32x4ShuffleMix:
 | 
| -    case MethodRecognizer::kFloat32x4Shuffle: {
 | 
| -      return InlineFloat32x4Getter(call, recognized_kind);
 | 
| -    }
 | 
| -    default:
 | 
| -      return false;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineFloat64x2Method(
 | 
| -    InstanceCallInstr* call,
 | 
| -    MethodRecognizer::Kind recognized_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(call->HasICData());
 | 
| -  switch (recognized_kind) {
 | 
| -    case MethodRecognizer::kFloat64x2GetX:
 | 
| -    case MethodRecognizer::kFloat64x2GetY:
 | 
| -      ASSERT(call->ic_data()->HasReceiverClassId(kFloat64x2Cid));
 | 
| -      ASSERT(call->ic_data()->HasOneTarget());
 | 
| -      return InlineFloat64x2Getter(call, recognized_kind);
 | 
| -    case MethodRecognizer::kFloat64x2Negate:
 | 
| -    case MethodRecognizer::kFloat64x2Abs:
 | 
| -    case MethodRecognizer::kFloat64x2Sqrt:
 | 
| -    case MethodRecognizer::kFloat64x2GetSignMask: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Float64x2ZeroArgInstr* zeroArg =
 | 
| -          new(I) Float64x2ZeroArgInstr(
 | 
| -              recognized_kind, new(I) Value(left), call->deopt_id());
 | 
| -      ReplaceCall(call, zeroArg);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kFloat64x2Scale:
 | 
| -    case MethodRecognizer::kFloat64x2WithX:
 | 
| -    case MethodRecognizer::kFloat64x2WithY:
 | 
| -    case MethodRecognizer::kFloat64x2Min:
 | 
| -    case MethodRecognizer::kFloat64x2Max: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      Definition* right = call->ArgumentAt(1);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Float64x2OneArgInstr* zeroArg =
 | 
| -          new(I) Float64x2OneArgInstr(recognized_kind,
 | 
| -                                      new(I) Value(left),
 | 
| -                                      new(I) Value(right),
 | 
| -                                      call->deopt_id());
 | 
| -      ReplaceCall(call, zeroArg);
 | 
| -      return true;
 | 
| -    }
 | 
| -    default:
 | 
| -      return false;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineInt32x4Method(
 | 
| -    InstanceCallInstr* call,
 | 
| -    MethodRecognizer::Kind recognized_kind) {
 | 
| -  if (!ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  ASSERT(call->HasICData());
 | 
| -  switch (recognized_kind) {
 | 
| -    case MethodRecognizer::kInt32x4ShuffleMix:
 | 
| -    case MethodRecognizer::kInt32x4Shuffle:
 | 
| -    case MethodRecognizer::kInt32x4GetFlagX:
 | 
| -    case MethodRecognizer::kInt32x4GetFlagY:
 | 
| -    case MethodRecognizer::kInt32x4GetFlagZ:
 | 
| -    case MethodRecognizer::kInt32x4GetFlagW:
 | 
| -    case MethodRecognizer::kInt32x4GetSignMask:
 | 
| -      ASSERT(call->ic_data()->HasReceiverClassId(kInt32x4Cid));
 | 
| -      ASSERT(call->ic_data()->HasOneTarget());
 | 
| -      return InlineInt32x4Getter(call, recognized_kind);
 | 
| -
 | 
| -    case MethodRecognizer::kInt32x4Select: {
 | 
| -      Definition* mask = call->ArgumentAt(0);
 | 
| -      Definition* trueValue = call->ArgumentAt(1);
 | 
| -      Definition* falseValue = call->ArgumentAt(2);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(mask,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Int32x4SelectInstr* select = new(I) Int32x4SelectInstr(
 | 
| -          new(I) Value(mask),
 | 
| -          new(I) Value(trueValue),
 | 
| -          new(I) Value(falseValue),
 | 
| -          call->deopt_id());
 | 
| -      ReplaceCall(call, select);
 | 
| -      return true;
 | 
| -    }
 | 
| -    case MethodRecognizer::kInt32x4WithFlagX:
 | 
| -    case MethodRecognizer::kInt32x4WithFlagY:
 | 
| -    case MethodRecognizer::kInt32x4WithFlagZ:
 | 
| -    case MethodRecognizer::kInt32x4WithFlagW: {
 | 
| -      Definition* left = call->ArgumentAt(0);
 | 
| -      Definition* flag = call->ArgumentAt(1);
 | 
| -      // Type check left.
 | 
| -      AddCheckClass(left,
 | 
| -                    ICData::ZoneHandle(
 | 
| -                        I, call->ic_data()->AsUnaryClassChecksForArgNr(0)),
 | 
| -                    call->deopt_id(),
 | 
| -                    call->env(),
 | 
| -                    call);
 | 
| -      Int32x4SetFlagInstr* setFlag = new(I) Int32x4SetFlagInstr(
 | 
| -          recognized_kind,
 | 
| -          new(I) Value(left),
 | 
| -          new(I) Value(flag),
 | 
| -          call->deopt_id());
 | 
| -      ReplaceCall(call, setFlag);
 | 
| -      return true;
 | 
| -    }
 | 
| -    default:
 | 
| -      return false;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineByteArrayViewLoad(Instruction* call,
 | 
| -                                                 Definition* receiver,
 | 
| -                                                 intptr_t array_cid,
 | 
| -                                                 intptr_t view_cid,
 | 
| -                                                 const ICData& ic_data,
 | 
| -                                                 TargetEntryInstr** entry,
 | 
| -                                                 Definition** last) {
 | 
| -  ASSERT(array_cid != kIllegalCid);
 | 
| -  Definition* array = receiver;
 | 
| -  Definition* index = call->ArgumentAt(1);
 | 
| -  *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(),
 | 
| -                                   call->GetBlock()->try_index());
 | 
| -  (*entry)->InheritDeoptTarget(I, call);
 | 
| -  Instruction* cursor = *entry;
 | 
| -
 | 
| -  array_cid = PrepareInlineByteArrayViewOp(call,
 | 
| -                                           array_cid,
 | 
| -                                           view_cid,
 | 
| -                                           &array,
 | 
| -                                           index,
 | 
| -                                           &cursor);
 | 
| -
 | 
| -  intptr_t deopt_id = Isolate::kNoDeoptId;
 | 
| -  if ((array_cid == kTypedDataInt32ArrayCid) ||
 | 
| -      (array_cid == kTypedDataUint32ArrayCid)) {
 | 
| -    // Deoptimization may be needed if result does not always fit in a Smi.
 | 
| -    deopt_id = (kSmiBits >= 32) ? Isolate::kNoDeoptId : call->deopt_id();
 | 
| -  }
 | 
| -
 | 
| -  *last = new(I) LoadIndexedInstr(new(I) Value(array),
 | 
| -                                  new(I) Value(index),
 | 
| -                                  1,
 | 
| -                                  view_cid,
 | 
| -                                  deopt_id,
 | 
| -                                  call->token_pos());
 | 
| -  cursor = flow_graph()->AppendTo(
 | 
| -      cursor,
 | 
| -      *last,
 | 
| -      deopt_id != Isolate::kNoDeoptId ? call->env() : NULL,
 | 
| -      FlowGraph::kValue);
 | 
| -
 | 
| -  if (view_cid == kTypedDataFloat32ArrayCid) {
 | 
| -    *last = new(I) FloatToDoubleInstr(new(I) Value(*last), deopt_id);
 | 
| -    flow_graph()->AppendTo(cursor,
 | 
| -                           *last,
 | 
| -                           deopt_id != Isolate::kNoDeoptId ? call->env() : NULL,
 | 
| -                           FlowGraph::kValue);
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::InlineByteArrayViewStore(const Function& target,
 | 
| -                                                  Instruction* call,
 | 
| -                                                  Definition* receiver,
 | 
| -                                                  intptr_t array_cid,
 | 
| -                                                  intptr_t view_cid,
 | 
| -                                                  const ICData& ic_data,
 | 
| -                                                  TargetEntryInstr** entry,
 | 
| -                                                  Definition** last) {
 | 
| -  ASSERT(array_cid != kIllegalCid);
 | 
| -  Definition* array = receiver;
 | 
| -  Definition* index = call->ArgumentAt(1);
 | 
| -  *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(),
 | 
| -                                   call->GetBlock()->try_index());
 | 
| -  (*entry)->InheritDeoptTarget(I, call);
 | 
| -  Instruction* cursor = *entry;
 | 
| -
 | 
| -  array_cid = PrepareInlineByteArrayViewOp(call,
 | 
| -                                           array_cid,
 | 
| -                                           view_cid,
 | 
| -                                           &array,
 | 
| -                                           index,
 | 
| -                                           &cursor);
 | 
| -
 | 
| -  // Extract the instance call so we can use the function_name in the stored
 | 
| -  // value check ICData.
 | 
| -  InstanceCallInstr* i_call = NULL;
 | 
| -  if (call->IsPolymorphicInstanceCall()) {
 | 
| -    i_call = call->AsPolymorphicInstanceCall()->instance_call();
 | 
| -  } else {
 | 
| -    ASSERT(call->IsInstanceCall());
 | 
| -    i_call = call->AsInstanceCall();
 | 
| -  }
 | 
| -  ASSERT(i_call != NULL);
 | 
| -  ICData& value_check = ICData::ZoneHandle(I);
 | 
| -  switch (view_cid) {
 | 
| -    case kTypedDataInt8ArrayCid:
 | 
| -    case kTypedDataUint8ArrayCid:
 | 
| -    case kTypedDataUint8ClampedArrayCid:
 | 
| -    case kExternalTypedDataUint8ArrayCid:
 | 
| -    case kExternalTypedDataUint8ClampedArrayCid:
 | 
| -    case kTypedDataInt16ArrayCid:
 | 
| -    case kTypedDataUint16ArrayCid: {
 | 
| -      // Check that value is always smi.
 | 
| -      value_check = ICData::New(flow_graph_->parsed_function()->function(),
 | 
| -                                i_call->function_name(),
 | 
| -                                Object::empty_array(),  // Dummy args. descr.
 | 
| -                                Isolate::kNoDeoptId,
 | 
| -                                1);
 | 
| -      value_check.AddReceiverCheck(kSmiCid, target);
 | 
| -      break;
 | 
| -    }
 | 
| -    case kTypedDataInt32ArrayCid:
 | 
| -    case kTypedDataUint32ArrayCid:
 | 
| -      // On 64-bit platforms assume that stored value is always a smi.
 | 
| -      if (kSmiBits >= 32) {
 | 
| -        value_check = ICData::New(flow_graph_->parsed_function()->function(),
 | 
| -                                  i_call->function_name(),
 | 
| -                                  Object::empty_array(),  // Dummy args. descr.
 | 
| -                                  Isolate::kNoDeoptId,
 | 
| -                                  1);
 | 
| -        value_check.AddReceiverCheck(kSmiCid, target);
 | 
| -      }
 | 
| -      break;
 | 
| -    case kTypedDataFloat32ArrayCid:
 | 
| -    case kTypedDataFloat64ArrayCid: {
 | 
| -      // Check that value is always double.
 | 
| -      value_check = ICData::New(flow_graph_->parsed_function()->function(),
 | 
| -                                i_call->function_name(),
 | 
| -                                Object::empty_array(),  // Dummy args. descr.
 | 
| -                                Isolate::kNoDeoptId,
 | 
| -                                1);
 | 
| -      value_check.AddReceiverCheck(kDoubleCid, target);
 | 
| -      break;
 | 
| -    }
 | 
| -    case kTypedDataInt32x4ArrayCid: {
 | 
| -      // Check that value is always Int32x4.
 | 
| -      value_check = ICData::New(flow_graph_->parsed_function()->function(),
 | 
| -                                i_call->function_name(),
 | 
| -                                Object::empty_array(),  // Dummy args. descr.
 | 
| -                                Isolate::kNoDeoptId,
 | 
| -                                1);
 | 
| -      value_check.AddReceiverCheck(kInt32x4Cid, target);
 | 
| -      break;
 | 
| -    }
 | 
| -    case kTypedDataFloat32x4ArrayCid: {
 | 
| -      // Check that value is always Float32x4.
 | 
| -      value_check = ICData::New(flow_graph_->parsed_function()->function(),
 | 
| -                                i_call->function_name(),
 | 
| -                                Object::empty_array(),  // Dummy args. descr.
 | 
| -                                Isolate::kNoDeoptId,
 | 
| -                                1);
 | 
| -      value_check.AddReceiverCheck(kFloat32x4Cid, target);
 | 
| -      break;
 | 
| -    }
 | 
| -    default:
 | 
| -      // Array cids are already checked in the caller.
 | 
| -      UNREACHABLE();
 | 
| -  }
 | 
| -
 | 
| -  Definition* stored_value = call->ArgumentAt(2);
 | 
| -  if (!value_check.IsNull()) {
 | 
| -    AddCheckClass(stored_value, value_check, call->deopt_id(), call->env(),
 | 
| -                  call);
 | 
| -  }
 | 
| -
 | 
| -  if (view_cid == kTypedDataFloat32ArrayCid) {
 | 
| -    stored_value = new(I) DoubleToFloatInstr(
 | 
| -        new(I) Value(stored_value), call->deopt_id());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    stored_value,
 | 
| -                                    NULL,
 | 
| -                                    FlowGraph::kValue);
 | 
| -  } else if (view_cid == kTypedDataInt32ArrayCid) {
 | 
| -    stored_value = new(I) UnboxInt32Instr(
 | 
| -        UnboxInt32Instr::kTruncate,
 | 
| -        new(I) Value(stored_value),
 | 
| -        call->deopt_id());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    stored_value,
 | 
| -                                    call->env(),
 | 
| -                                    FlowGraph::kValue);
 | 
| -  } else if (view_cid == kTypedDataUint32ArrayCid) {
 | 
| -    stored_value = new(I) UnboxUint32Instr(
 | 
| -        new(I) Value(stored_value),
 | 
| -        call->deopt_id());
 | 
| -    ASSERT(stored_value->AsUnboxInteger()->is_truncating());
 | 
| -    cursor = flow_graph()->AppendTo(cursor,
 | 
| -                                    stored_value,
 | 
| -                                    call->env(),
 | 
| -                                    FlowGraph::kValue);
 | 
| -  }
 | 
| -
 | 
| -  StoreBarrierType needs_store_barrier = kNoStoreBarrier;
 | 
| -  *last = new(I) StoreIndexedInstr(new(I) Value(array),
 | 
| -                                   new(I) Value(index),
 | 
| -                                   new(I) Value(stored_value),
 | 
| -                                   needs_store_barrier,
 | 
| -                                   1,  // Index scale
 | 
| -                                   view_cid,
 | 
| -                                   call->deopt_id(),
 | 
| -                                   call->token_pos());
 | 
| -
 | 
| -  flow_graph()->AppendTo(cursor,
 | 
| -                         *last,
 | 
| -                         call->deopt_id() != Isolate::kNoDeoptId ?
 | 
| -                            call->env() : NULL,
 | 
| -                         FlowGraph::kEffect);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -
 | 
| -intptr_t FlowGraphOptimizer::PrepareInlineByteArrayViewOp(
 | 
| -    Instruction* call,
 | 
| -    intptr_t array_cid,
 | 
| -    intptr_t view_cid,
 | 
| -    Definition** array,
 | 
| -    Definition* byte_index,
 | 
| -    Instruction** cursor) {
 | 
| -  // Insert byte_index smi check.
 | 
| -  *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                   new(I) CheckSmiInstr(
 | 
| -                                       new(I) Value(byte_index),
 | 
| -                                       call->deopt_id(),
 | 
| -                                       call->token_pos()),
 | 
| -                                   call->env(),
 | 
| -                                   FlowGraph::kEffect);
 | 
| -
 | 
| -  LoadFieldInstr* length =
 | 
| -      new(I) LoadFieldInstr(
 | 
| -          new(I) Value(*array),
 | 
| -          CheckArrayBoundInstr::LengthOffsetFor(array_cid),
 | 
| -          Type::ZoneHandle(I, Type::SmiType()),
 | 
| -          call->token_pos());
 | 
| -  length->set_is_immutable(true);
 | 
| -  length->set_result_cid(kSmiCid);
 | 
| -  length->set_recognized_kind(
 | 
| -      LoadFieldInstr::RecognizedKindFromArrayCid(array_cid));
 | 
| -  *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                   length,
 | 
| -                                   NULL,
 | 
| -                                   FlowGraph::kValue);
 | 
| -
 | 
| -  intptr_t element_size = Instance::ElementSizeFor(array_cid);
 | 
| -  ConstantInstr* bytes_per_element =
 | 
| -      flow_graph()->GetConstant(Smi::Handle(I, Smi::New(element_size)));
 | 
| -  BinarySmiOpInstr* len_in_bytes =
 | 
| -      new(I) BinarySmiOpInstr(Token::kMUL,
 | 
| -                              new(I) Value(length),
 | 
| -                              new(I) Value(bytes_per_element),
 | 
| -                              call->deopt_id());
 | 
| -  *cursor = flow_graph()->AppendTo(*cursor, len_in_bytes, call->env(),
 | 
| -                                   FlowGraph::kValue);
 | 
| -
 | 
| -  // adjusted_length = len_in_bytes - (element_size - 1).
 | 
| -  Definition* adjusted_length = len_in_bytes;
 | 
| -  intptr_t adjustment = Instance::ElementSizeFor(view_cid) - 1;
 | 
| -  if (adjustment > 0) {
 | 
| -    ConstantInstr* length_adjustment =
 | 
| -        flow_graph()->GetConstant(Smi::Handle(I, Smi::New(adjustment)));
 | 
| -    adjusted_length =
 | 
| -        new(I) BinarySmiOpInstr(Token::kSUB,
 | 
| -                                new(I) Value(len_in_bytes),
 | 
| -                                new(I) Value(length_adjustment),
 | 
| -                                call->deopt_id());
 | 
| -    *cursor = flow_graph()->AppendTo(*cursor, adjusted_length, call->env(),
 | 
| -                                     FlowGraph::kValue);
 | 
| -  }
 | 
| -
 | 
| -  // Check adjusted_length > 0.
 | 
| -  ConstantInstr* zero =
 | 
| -      flow_graph()->GetConstant(Smi::Handle(I, Smi::New(0)));
 | 
| -  *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                   new(I) CheckArrayBoundInstr(
 | 
| -                                       new(I) Value(adjusted_length),
 | 
| -                                       new(I) Value(zero),
 | 
| -                                       call->deopt_id()),
 | 
| -                                   call->env(),
 | 
| -                                   FlowGraph::kEffect);
 | 
| -  // Check 0 <= byte_index < adjusted_length.
 | 
| -  *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                   new(I) CheckArrayBoundInstr(
 | 
| -                                       new(I) Value(adjusted_length),
 | 
| -                                       new(I) Value(byte_index),
 | 
| -                                       call->deopt_id()),
 | 
| -                                   call->env(),
 | 
| -                                   FlowGraph::kEffect);
 | 
| -
 | 
| -  if (RawObject::IsExternalTypedDataClassId(array_cid)) {
 | 
| -    LoadUntaggedInstr* elements =
 | 
| -        new(I) LoadUntaggedInstr(new(I) Value(*array),
 | 
| -                                 ExternalTypedData::data_offset());
 | 
| -    *cursor = flow_graph()->AppendTo(*cursor,
 | 
| -                                     elements,
 | 
| -                                     NULL,
 | 
| -                                     FlowGraph::kValue);
 | 
| -    *array = elements;
 | 
| -  }
 | 
| -  return array_cid;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::BuildByteArrayViewLoad(InstanceCallInstr* call,
 | 
| -                                                intptr_t view_cid) {
 | 
| -  const bool simd_view = (view_cid == kTypedDataFloat32x4ArrayCid) ||
 | 
| -                         (view_cid == kTypedDataInt32x4ArrayCid);
 | 
| -  const bool float_view = (view_cid == kTypedDataFloat32ArrayCid) ||
 | 
| -                          (view_cid == kTypedDataFloat64ArrayCid);
 | 
| -  if (float_view && !CanUnboxDouble()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  if (simd_view && !ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  return TryReplaceInstanceCallWithInline(call);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::BuildByteArrayViewStore(InstanceCallInstr* call,
 | 
| -                                                 intptr_t view_cid) {
 | 
| -  const bool simd_view = (view_cid == kTypedDataFloat32x4ArrayCid) ||
 | 
| -                         (view_cid == kTypedDataInt32x4ArrayCid);
 | 
| -  const bool float_view = (view_cid == kTypedDataFloat32ArrayCid) ||
 | 
| -                          (view_cid == kTypedDataFloat64ArrayCid);
 | 
| -  if (float_view && !CanUnboxDouble()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  if (simd_view && !ShouldInlineSimd()) {
 | 
| -    return false;
 | 
| -  }
 | 
| -  return TryReplaceInstanceCallWithInline(call);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// If type tests specified by 'ic_data' do not depend on type arguments,
 | 
| -// return mapping cid->result in 'results' (i : cid; i + 1: result).
 | 
| -// If all tests yield the same result, return it otherwise return Bool::null.
 | 
| -// If no mapping is possible, 'results' is empty.
 | 
| -// An instance-of test returning all same results can be converted to a class
 | 
| -// check.
 | 
| -RawBool* FlowGraphOptimizer::InstanceOfAsBool(
 | 
| -    const ICData& ic_data,
 | 
| -    const AbstractType& type,
 | 
| -    ZoneGrowableArray<intptr_t>* results) const {
 | 
| -  ASSERT(results->is_empty());
 | 
| -  ASSERT(ic_data.NumArgsTested() == 1);  // Unary checks only.
 | 
| -  if (!type.IsInstantiated() || type.IsMalformedOrMalbounded()) {
 | 
| -    return Bool::null();
 | 
| -  }
 | 
| -  const Class& type_class = Class::Handle(I, type.type_class());
 | 
| -  const intptr_t num_type_args = type_class.NumTypeArguments();
 | 
| -  if (num_type_args > 0) {
 | 
| -    // Only raw types can be directly compared, thus disregarding type
 | 
| -    // arguments.
 | 
| -    const intptr_t num_type_params = type_class.NumTypeParameters();
 | 
| -    const intptr_t from_index = num_type_args - num_type_params;
 | 
| -    const TypeArguments& type_arguments =
 | 
| -        TypeArguments::Handle(I, type.arguments());
 | 
| -    const bool is_raw_type = type_arguments.IsNull() ||
 | 
| -        type_arguments.IsRaw(from_index, num_type_params);
 | 
| -    if (!is_raw_type) {
 | 
| -      // Unknown result.
 | 
| -      return Bool::null();
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  const ClassTable& class_table = *isolate()->class_table();
 | 
| -  Bool& prev = Bool::Handle(I);
 | 
| -  Class& cls = Class::Handle(I);
 | 
| -
 | 
| -  bool results_differ = false;
 | 
| -  for (int i = 0; i < ic_data.NumberOfChecks(); i++) {
 | 
| -    cls = class_table.At(ic_data.GetReceiverClassIdAt(i));
 | 
| -    if (cls.NumTypeArguments() > 0) {
 | 
| -      return Bool::null();
 | 
| -    }
 | 
| -    const bool is_subtype = cls.IsSubtypeOf(
 | 
| -        TypeArguments::Handle(I),
 | 
| -        type_class,
 | 
| -        TypeArguments::Handle(I),
 | 
| -        NULL);
 | 
| -    results->Add(cls.id());
 | 
| -    results->Add(is_subtype);
 | 
| -    if (prev.IsNull()) {
 | 
| -      prev = Bool::Get(is_subtype).raw();
 | 
| -    } else {
 | 
| -      if (is_subtype != prev.value()) {
 | 
| -        results_differ = true;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -  return results_differ ?  Bool::null() : prev.raw();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Returns true if checking against this type is a direct class id comparison.
 | 
| -bool FlowGraphOptimizer::TypeCheckAsClassEquality(const AbstractType& type) {
 | 
| -  ASSERT(type.IsFinalized() && !type.IsMalformedOrMalbounded());
 | 
| -  // Requires CHA.
 | 
| -  if (!FLAG_use_cha) return false;
 | 
| -  if (!type.IsInstantiated()) return false;
 | 
| -  const Class& type_class = Class::Handle(type.type_class());
 | 
| -  // Signature classes have different type checking rules.
 | 
| -  if (type_class.IsSignatureClass()) return false;
 | 
| -  // Could be an interface check?
 | 
| -  if (isolate()->cha()->IsImplemented(type_class)) return false;
 | 
| -  // Check if there are subclasses.
 | 
| -  if (isolate()->cha()->HasSubclasses(type_class)) return false;
 | 
| -  const intptr_t num_type_args = type_class.NumTypeArguments();
 | 
| -  if (num_type_args > 0) {
 | 
| -    // Only raw types can be directly compared, thus disregarding type
 | 
| -    // arguments.
 | 
| -    const intptr_t num_type_params = type_class.NumTypeParameters();
 | 
| -    const intptr_t from_index = num_type_args - num_type_params;
 | 
| -    const TypeArguments& type_arguments =
 | 
| -        TypeArguments::Handle(type.arguments());
 | 
| -    const bool is_raw_type = type_arguments.IsNull() ||
 | 
| -        type_arguments.IsRaw(from_index, num_type_params);
 | 
| -    return is_raw_type;
 | 
| -  }
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool CidTestResultsContains(const ZoneGrowableArray<intptr_t>& results,
 | 
| -                                   intptr_t test_cid) {
 | 
| -  for (intptr_t i = 0; i < results.length(); i += 2) {
 | 
| -    if (results[i] == test_cid) return true;
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static void TryAddTest(ZoneGrowableArray<intptr_t>* results,
 | 
| -                       intptr_t test_cid,
 | 
| -                       bool result) {
 | 
| -  if (!CidTestResultsContains(*results, test_cid)) {
 | 
| -    results->Add(test_cid);
 | 
| -    results->Add(result);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Tries to add cid tests to 'results' so that no deoptimization is
 | 
| -// necessary.
 | 
| -// TODO(srdjan): Do also for other than 'int' type.
 | 
| -static bool TryExpandTestCidsResult(ZoneGrowableArray<intptr_t>* results,
 | 
| -                                    const AbstractType& type) {
 | 
| -  ASSERT(results->length() >= 2);  // At least on eentry.
 | 
| -  const ClassTable& class_table = *Isolate::Current()->class_table();
 | 
| -  if ((*results)[0] != kSmiCid) {
 | 
| -    const Class& cls = Class::Handle(class_table.At(kSmiCid));
 | 
| -    const Class& type_class = Class::Handle(type.type_class());
 | 
| -    const bool smi_is_subtype = cls.IsSubtypeOf(TypeArguments::Handle(),
 | 
| -                                                type_class,
 | 
| -                                                TypeArguments::Handle(),
 | 
| -                                                NULL);
 | 
| -    results->Add((*results)[results->length() - 2]);
 | 
| -    results->Add((*results)[results->length() - 2]);
 | 
| -    for (intptr_t i = results->length() - 3; i > 1; --i) {
 | 
| -      (*results)[i] = (*results)[i - 2];
 | 
| -    }
 | 
| -    (*results)[0] = kSmiCid;
 | 
| -    (*results)[1] = smi_is_subtype;
 | 
| -  }
 | 
| -
 | 
| -  ASSERT(type.IsInstantiated() && !type.IsMalformedOrMalbounded());
 | 
| -  ASSERT(results->length() >= 2);
 | 
| -  if (type.IsIntType()) {
 | 
| -    ASSERT((*results)[0] == kSmiCid);
 | 
| -    TryAddTest(results, kMintCid, true);
 | 
| -    TryAddTest(results, kBigintCid, true);
 | 
| -    // Cannot deoptimize since all tests returning true have been added.
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  return true;  // May deoptimize since we have not identified all 'true' tests.
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// TODO(srdjan): Use ICData to check if always true or false.
 | 
| -void FlowGraphOptimizer::ReplaceWithInstanceOf(InstanceCallInstr* call) {
 | 
| -  ASSERT(Token::IsTypeTestOperator(call->token_kind()));
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* instantiator = call->ArgumentAt(1);
 | 
| -  Definition* type_args = call->ArgumentAt(2);
 | 
| -  const AbstractType& type =
 | 
| -      AbstractType::Cast(call->ArgumentAt(3)->AsConstant()->value());
 | 
| -  const bool negate = Bool::Cast(
 | 
| -      call->ArgumentAt(4)->OriginalDefinition()->AsConstant()->value()).value();
 | 
| -  const ICData& unary_checks =
 | 
| -      ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks());
 | 
| -  if (FLAG_warn_on_javascript_compatibility &&
 | 
| -      !unary_checks.IssuedJSWarning() &&
 | 
| -      (type.IsIntType() || type.IsDoubleType() || !type.IsInstantiated())) {
 | 
| -    // No warning was reported yet for this type check, either because it has
 | 
| -    // not been executed yet, or because no problematic combinations of instance
 | 
| -    // type and test type have been encountered so far. A warning may still be
 | 
| -    // reported, so do not replace the instance call.
 | 
| -    return;
 | 
| -  }
 | 
| -  if (unary_checks.NumberOfChecks() <= FLAG_max_polymorphic_checks) {
 | 
| -    ZoneGrowableArray<intptr_t>* results =
 | 
| -        new(I) ZoneGrowableArray<intptr_t>(unary_checks.NumberOfChecks() * 2);
 | 
| -    Bool& as_bool =
 | 
| -        Bool::ZoneHandle(I, InstanceOfAsBool(unary_checks, type, results));
 | 
| -    if (as_bool.IsNull()) {
 | 
| -      if (results->length() == unary_checks.NumberOfChecks() * 2) {
 | 
| -        const bool can_deopt = TryExpandTestCidsResult(results, type);
 | 
| -        TestCidsInstr* test_cids = new(I) TestCidsInstr(
 | 
| -            call->token_pos(),
 | 
| -            negate ? Token::kISNOT : Token::kIS,
 | 
| -            new(I) Value(left),
 | 
| -            *results,
 | 
| -            can_deopt ? call->deopt_id() : Isolate::kNoDeoptId);
 | 
| -        // Remove type.
 | 
| -        ReplaceCall(call, test_cids);
 | 
| -        return;
 | 
| -      }
 | 
| -    } else {
 | 
| -      // TODO(srdjan): Use TestCidsInstr also for this case.
 | 
| -      // One result only.
 | 
| -      AddReceiverCheck(call);
 | 
| -      if (negate) {
 | 
| -        as_bool = Bool::Get(!as_bool.value()).raw();
 | 
| -      }
 | 
| -      ConstantInstr* bool_const = flow_graph()->GetConstant(as_bool);
 | 
| -      for (intptr_t i = 0; i < call->ArgumentCount(); ++i) {
 | 
| -        PushArgumentInstr* push = call->PushArgumentAt(i);
 | 
| -        push->ReplaceUsesWith(push->value()->definition());
 | 
| -        push->RemoveFromGraph();
 | 
| -      }
 | 
| -      call->ReplaceUsesWith(bool_const);
 | 
| -      ASSERT(current_iterator()->Current() == call);
 | 
| -      current_iterator()->RemoveCurrentFromGraph();
 | 
| -      return;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (TypeCheckAsClassEquality(type)) {
 | 
| -    LoadClassIdInstr* left_cid = new(I) LoadClassIdInstr(new(I) Value(left));
 | 
| -    InsertBefore(call,
 | 
| -                 left_cid,
 | 
| -                 NULL,
 | 
| -                 FlowGraph::kValue);
 | 
| -    const intptr_t type_cid = Class::Handle(I, type.type_class()).id();
 | 
| -    ConstantInstr* cid =
 | 
| -        flow_graph()->GetConstant(Smi::Handle(I, Smi::New(type_cid)));
 | 
| -
 | 
| -    StrictCompareInstr* check_cid =
 | 
| -        new(I) StrictCompareInstr(
 | 
| -            call->token_pos(),
 | 
| -            negate ? Token::kNE_STRICT : Token::kEQ_STRICT,
 | 
| -            new(I) Value(left_cid),
 | 
| -            new(I) Value(cid),
 | 
| -            false);  // No number check.
 | 
| -    ReplaceCall(call, check_cid);
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  InstanceOfInstr* instance_of =
 | 
| -      new(I) InstanceOfInstr(call->token_pos(),
 | 
| -                             new(I) Value(left),
 | 
| -                             new(I) Value(instantiator),
 | 
| -                             new(I) Value(type_args),
 | 
| -                             type,
 | 
| -                             negate,
 | 
| -                             call->deopt_id());
 | 
| -  ReplaceCall(call, instance_of);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// TODO(srdjan): Apply optimizations as in ReplaceWithInstanceOf (TestCids).
 | 
| -void FlowGraphOptimizer::ReplaceWithTypeCast(InstanceCallInstr* call) {
 | 
| -  ASSERT(Token::IsTypeCastOperator(call->token_kind()));
 | 
| -  Definition* left = call->ArgumentAt(0);
 | 
| -  Definition* instantiator = call->ArgumentAt(1);
 | 
| -  Definition* type_args = call->ArgumentAt(2);
 | 
| -  const AbstractType& type =
 | 
| -      AbstractType::Cast(call->ArgumentAt(3)->AsConstant()->value());
 | 
| -  ASSERT(!type.IsMalformedOrMalbounded());
 | 
| -  const ICData& unary_checks =
 | 
| -      ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks());
 | 
| -  if (FLAG_warn_on_javascript_compatibility &&
 | 
| -      !unary_checks.IssuedJSWarning() &&
 | 
| -      (type.IsIntType() || type.IsDoubleType() || !type.IsInstantiated())) {
 | 
| -    // No warning was reported yet for this type check, either because it has
 | 
| -    // not been executed yet, or because no problematic combinations of instance
 | 
| -    // type and test type have been encountered so far. A warning may still be
 | 
| -    // reported, so do not replace the instance call.
 | 
| -    return;
 | 
| -  }
 | 
| -  if (unary_checks.NumberOfChecks() <= FLAG_max_polymorphic_checks) {
 | 
| -    ZoneGrowableArray<intptr_t>* results =
 | 
| -        new(I) ZoneGrowableArray<intptr_t>(unary_checks.NumberOfChecks() * 2);
 | 
| -    const Bool& as_bool = Bool::ZoneHandle(I,
 | 
| -        InstanceOfAsBool(unary_checks, type, results));
 | 
| -    if (as_bool.raw() == Bool::True().raw()) {
 | 
| -      AddReceiverCheck(call);
 | 
| -      // Remove the original push arguments.
 | 
| -      for (intptr_t i = 0; i < call->ArgumentCount(); ++i) {
 | 
| -        PushArgumentInstr* push = call->PushArgumentAt(i);
 | 
| -        push->ReplaceUsesWith(push->value()->definition());
 | 
| -        push->RemoveFromGraph();
 | 
| -      }
 | 
| -      // Remove call, replace it with 'left'.
 | 
| -      call->ReplaceUsesWith(left);
 | 
| -      ASSERT(current_iterator()->Current() == call);
 | 
| -      current_iterator()->RemoveCurrentFromGraph();
 | 
| -      return;
 | 
| -    }
 | 
| -  }
 | 
| -  const String& dst_name = String::ZoneHandle(I,
 | 
| -      Symbols::New(Exceptions::kCastErrorDstName));
 | 
| -  AssertAssignableInstr* assert_as =
 | 
| -      new(I) AssertAssignableInstr(call->token_pos(),
 | 
| -                                   new(I) Value(left),
 | 
| -                                   new(I) Value(instantiator),
 | 
| -                                   new(I) Value(type_args),
 | 
| -                                   type,
 | 
| -                                   dst_name,
 | 
| -                                   call->deopt_id());
 | 
| -  ReplaceCall(call, assert_as);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Tries to optimize instance call by replacing it with a faster instruction
 | 
| -// (e.g, binary op, field load, ..).
 | 
| -void FlowGraphOptimizer::VisitInstanceCall(InstanceCallInstr* instr) {
 | 
| -  if (!instr->HasICData() || (instr->ic_data()->NumberOfUsedChecks() == 0)) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  const Token::Kind op_kind = instr->token_kind();
 | 
| -  // Type test is special as it always gets converted into inlined code.
 | 
| -  if (Token::IsTypeTestOperator(op_kind)) {
 | 
| -    ReplaceWithInstanceOf(instr);
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  if (Token::IsTypeCastOperator(op_kind)) {
 | 
| -    ReplaceWithTypeCast(instr);
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  const ICData& unary_checks =
 | 
| -      ICData::ZoneHandle(I, instr->ic_data()->AsUnaryClassChecks());
 | 
| -
 | 
| -  const intptr_t max_checks = (op_kind == Token::kEQ)
 | 
| -      ? FLAG_max_equality_polymorphic_checks
 | 
| -      : FLAG_max_polymorphic_checks;
 | 
| -  if ((unary_checks.NumberOfChecks() > max_checks) &&
 | 
| -      InstanceCallNeedsClassCheck(instr, RawFunction::kRegularFunction)) {
 | 
| -    // Too many checks, it will be megamorphic which needs unary checks.
 | 
| -    instr->set_ic_data(&unary_checks);
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  if ((op_kind == Token::kASSIGN_INDEX) && TryReplaceWithStoreIndexed(instr)) {
 | 
| -    return;
 | 
| -  }
 | 
| -  if ((op_kind == Token::kINDEX) && TryReplaceWithLoadIndexed(instr)) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  if (op_kind == Token::kEQ && TryReplaceWithEqualityOp(instr, op_kind)) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  if (Token::IsRelationalOperator(op_kind) &&
 | 
| -      TryReplaceWithRelationalOp(instr, op_kind)) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  if (Token::IsBinaryOperator(op_kind) &&
 | 
| -      TryReplaceWithBinaryOp(instr, op_kind)) {
 | 
| -    return;
 | 
| -  }
 | 
| -  if (Token::IsUnaryOperator(op_kind) &&
 | 
| -      TryReplaceWithUnaryOp(instr, op_kind)) {
 | 
| -    return;
 | 
| -  }
 | 
| -  if ((op_kind == Token::kGET) && TryInlineInstanceGetter(instr)) {
 | 
| -    return;
 | 
| -  }
 | 
| -  if ((op_kind == Token::kSET) &&
 | 
| -      TryInlineInstanceSetter(instr, unary_checks)) {
 | 
| -    return;
 | 
| -  }
 | 
| -  if (TryInlineInstanceMethod(instr)) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  bool has_one_target = unary_checks.HasOneTarget();
 | 
| -
 | 
| -  if (has_one_target) {
 | 
| -    // Check if the single target is a polymorphic target, if it is,
 | 
| -    // we don't have one target.
 | 
| -    const Function& target =
 | 
| -        Function::Handle(I, unary_checks.GetTargetAt(0));
 | 
| -    const bool polymorphic_target = MethodRecognizer::PolymorphicTarget(target);
 | 
| -    has_one_target = !polymorphic_target;
 | 
| -  }
 | 
| -
 | 
| -  if (has_one_target) {
 | 
| -    RawFunction::Kind function_kind =
 | 
| -        Function::Handle(I, unary_checks.GetTargetAt(0)).kind();
 | 
| -    if (!InstanceCallNeedsClassCheck(instr, function_kind)) {
 | 
| -      const bool call_with_checks = false;
 | 
| -      PolymorphicInstanceCallInstr* call =
 | 
| -          new(I) PolymorphicInstanceCallInstr(instr, unary_checks,
 | 
| -                                              call_with_checks);
 | 
| -      instr->ReplaceWith(call, current_iterator());
 | 
| -      return;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (unary_checks.NumberOfChecks() <= FLAG_max_polymorphic_checks) {
 | 
| -    bool call_with_checks;
 | 
| -    if (has_one_target) {
 | 
| -      // Type propagation has not run yet, we cannot eliminate the check.
 | 
| -      AddReceiverCheck(instr);
 | 
| -      // Call can still deoptimize, do not detach environment from instr.
 | 
| -      call_with_checks = false;
 | 
| -    } else {
 | 
| -      call_with_checks = true;
 | 
| -    }
 | 
| -    PolymorphicInstanceCallInstr* call =
 | 
| -        new(I) PolymorphicInstanceCallInstr(instr, unary_checks,
 | 
| -                                            call_with_checks);
 | 
| -    instr->ReplaceWith(call, current_iterator());
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::VisitStaticCall(StaticCallInstr* call) {
 | 
| -  if (!CanUnboxDouble()) {
 | 
| -    return;
 | 
| -  }
 | 
| -  MethodRecognizer::Kind recognized_kind =
 | 
| -      MethodRecognizer::RecognizeKind(call->function());
 | 
| -  MathUnaryInstr::MathUnaryKind unary_kind;
 | 
| -  switch (recognized_kind) {
 | 
| -    case MethodRecognizer::kMathSqrt:
 | 
| -      unary_kind = MathUnaryInstr::kSqrt;
 | 
| -      break;
 | 
| -    case MethodRecognizer::kMathSin:
 | 
| -      unary_kind = MathUnaryInstr::kSin;
 | 
| -      break;
 | 
| -    case MethodRecognizer::kMathCos:
 | 
| -      unary_kind = MathUnaryInstr::kCos;
 | 
| -      break;
 | 
| -    default:
 | 
| -      unary_kind = MathUnaryInstr::kIllegal;
 | 
| -      break;
 | 
| -  }
 | 
| -  if (unary_kind != MathUnaryInstr::kIllegal) {
 | 
| -    MathUnaryInstr* math_unary =
 | 
| -        new(I) MathUnaryInstr(unary_kind,
 | 
| -                              new(I) Value(call->ArgumentAt(0)),
 | 
| -                              call->deopt_id());
 | 
| -    ReplaceCall(call, math_unary);
 | 
| -  } else if ((recognized_kind == MethodRecognizer::kFloat32x4Zero) ||
 | 
| -             (recognized_kind == MethodRecognizer::kFloat32x4Splat) ||
 | 
| -             (recognized_kind == MethodRecognizer::kFloat32x4Constructor) ||
 | 
| -             (recognized_kind == MethodRecognizer::kFloat32x4FromFloat64x2)) {
 | 
| -    TryInlineFloat32x4Constructor(call, recognized_kind);
 | 
| -  } else if ((recognized_kind == MethodRecognizer::kFloat64x2Constructor) ||
 | 
| -             (recognized_kind == MethodRecognizer::kFloat64x2Zero) ||
 | 
| -             (recognized_kind == MethodRecognizer::kFloat64x2Splat) ||
 | 
| -             (recognized_kind == MethodRecognizer::kFloat64x2FromFloat32x4)) {
 | 
| -    TryInlineFloat64x2Constructor(call, recognized_kind);
 | 
| -  } else if ((recognized_kind == MethodRecognizer::kInt32x4BoolConstructor) ||
 | 
| -             (recognized_kind == MethodRecognizer::kInt32x4Constructor)) {
 | 
| -    TryInlineInt32x4Constructor(call, recognized_kind);
 | 
| -  } else if (recognized_kind == MethodRecognizer::kObjectConstructor) {
 | 
| -    // Remove the original push arguments.
 | 
| -    for (intptr_t i = 0; i < call->ArgumentCount(); ++i) {
 | 
| -      PushArgumentInstr* push = call->PushArgumentAt(i);
 | 
| -      push->ReplaceUsesWith(push->value()->definition());
 | 
| -      push->RemoveFromGraph();
 | 
| -    }
 | 
| -    // Manually replace call with global null constant. ReplaceCall can't
 | 
| -    // be used for definitions that are already in the graph.
 | 
| -    call->ReplaceUsesWith(flow_graph_->constant_null());
 | 
| -    ASSERT(current_iterator()->Current() == call);
 | 
| -    current_iterator()->RemoveCurrentFromGraph();;
 | 
| -  } else if ((recognized_kind == MethodRecognizer::kMathMin) ||
 | 
| -             (recognized_kind == MethodRecognizer::kMathMax)) {
 | 
| -    // We can handle only monomorphic min/max call sites with both arguments
 | 
| -    // being either doubles or smis.
 | 
| -    if (call->HasICData() && (call->ic_data()->NumberOfChecks() == 1)) {
 | 
| -      const ICData& ic_data = *call->ic_data();
 | 
| -      intptr_t result_cid = kIllegalCid;
 | 
| -      if (ICDataHasReceiverArgumentClassIds(ic_data, kDoubleCid, kDoubleCid)) {
 | 
| -        result_cid = kDoubleCid;
 | 
| -      } else if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) {
 | 
| -        result_cid = kSmiCid;
 | 
| -      }
 | 
| -      if (result_cid != kIllegalCid) {
 | 
| -        MathMinMaxInstr* min_max = new(I) MathMinMaxInstr(
 | 
| -            recognized_kind,
 | 
| -            new(I) Value(call->ArgumentAt(0)),
 | 
| -            new(I) Value(call->ArgumentAt(1)),
 | 
| -            call->deopt_id(),
 | 
| -            result_cid);
 | 
| -        const ICData& unary_checks =
 | 
| -            ICData::ZoneHandle(I, ic_data.AsUnaryClassChecks());
 | 
| -        AddCheckClass(min_max->left()->definition(),
 | 
| -                      unary_checks,
 | 
| -                      call->deopt_id(),
 | 
| -                      call->env(),
 | 
| -                      call);
 | 
| -        AddCheckClass(min_max->right()->definition(),
 | 
| -                      unary_checks,
 | 
| -                      call->deopt_id(),
 | 
| -                      call->env(),
 | 
| -                      call);
 | 
| -        ReplaceCall(call, min_max);
 | 
| -      }
 | 
| -    }
 | 
| -  } else if (recognized_kind == MethodRecognizer::kMathDoublePow) {
 | 
| -    // We know that first argument is double, the second is num.
 | 
| -    // InvokeMathCFunctionInstr requires unboxed doubles. UnboxDouble
 | 
| -    // instructions contain type checks and conversions to double.
 | 
| -    ZoneGrowableArray<Value*>* args =
 | 
| -        new(I) ZoneGrowableArray<Value*>(call->ArgumentCount());
 | 
| -    for (intptr_t i = 0; i < call->ArgumentCount(); i++) {
 | 
| -      args->Add(new(I) Value(call->ArgumentAt(i)));
 | 
| -    }
 | 
| -    InvokeMathCFunctionInstr* invoke =
 | 
| -        new(I) InvokeMathCFunctionInstr(args,
 | 
| -                                        call->deopt_id(),
 | 
| -                                        recognized_kind,
 | 
| -                                        call->token_pos());
 | 
| -    ReplaceCall(call, invoke);
 | 
| -  } else if (recognized_kind == MethodRecognizer::kDoubleFromInteger) {
 | 
| -    if (call->HasICData() && (call->ic_data()->NumberOfChecks() == 1)) {
 | 
| -      const ICData& ic_data = *call->ic_data();
 | 
| -      if (CanUnboxDouble()) {
 | 
| -        if (ArgIsAlways(kSmiCid, ic_data, 1)) {
 | 
| -          Definition* arg = call->ArgumentAt(1);
 | 
| -          AddCheckSmi(arg, call->deopt_id(), call->env(), call);
 | 
| -          ReplaceCall(call,
 | 
| -                      new(I) SmiToDoubleInstr(new(I) Value(arg),
 | 
| -                                              call->token_pos()));
 | 
| -        } else if (ArgIsAlways(kMintCid, ic_data, 1) &&
 | 
| -                   CanConvertUnboxedMintToDouble()) {
 | 
| -          Definition* arg = call->ArgumentAt(1);
 | 
| -          ReplaceCall(call,
 | 
| -                      new(I) MintToDoubleInstr(new(I) Value(arg),
 | 
| -                                               call->deopt_id()));
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  } else if (call->function().IsFactory()) {
 | 
| -    const Class& function_class =
 | 
| -        Class::Handle(I, call->function().Owner());
 | 
| -    if ((function_class.library() == Library::CoreLibrary()) ||
 | 
| -        (function_class.library() == Library::TypedDataLibrary())) {
 | 
| -      intptr_t cid = FactoryRecognizer::ResultCid(call->function());
 | 
| -      switch (cid) {
 | 
| -        case kArrayCid: {
 | 
| -          Value* type = new(I) Value(call->ArgumentAt(0));
 | 
| -          Value* num_elements = new(I) Value(call->ArgumentAt(1));
 | 
| -          if (num_elements->BindsToConstant() &&
 | 
| -              num_elements->BoundConstant().IsSmi()) {
 | 
| -            intptr_t length = Smi::Cast(num_elements->BoundConstant()).Value();
 | 
| -            if (length >= 0 && length <= Array::kMaxElements) {
 | 
| -              CreateArrayInstr* create_array =
 | 
| -                  new(I) CreateArrayInstr(
 | 
| -                      call->token_pos(), type, num_elements);
 | 
| -              ReplaceCall(call, create_array);
 | 
| -            }
 | 
| -          }
 | 
| -        }
 | 
| -        default:
 | 
| -          break;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::VisitStoreInstanceField(
 | 
| -    StoreInstanceFieldInstr* instr) {
 | 
| -  if (instr->IsUnboxedStore()) {
 | 
| -    ASSERT(instr->is_initialization_);
 | 
| -    // Determine if this field should be unboxed based on the usage of getter
 | 
| -    // and setter functions: The heuristic requires that the setter has a
 | 
| -    // usage count of at least 1/kGetterSetterRatio of the getter usage count.
 | 
| -    // This is to avoid unboxing fields where the setter is never or rarely
 | 
| -    // executed.
 | 
| -    const Field& field = Field::ZoneHandle(I, instr->field().raw());
 | 
| -    const String& field_name = String::Handle(I, field.name());
 | 
| -    const Class& owner = Class::Handle(I, field.owner());
 | 
| -    const Function& getter =
 | 
| -        Function::Handle(I, owner.LookupGetterFunction(field_name));
 | 
| -    const Function& setter =
 | 
| -        Function::Handle(I, owner.LookupSetterFunction(field_name));
 | 
| -    bool result = !getter.IsNull()
 | 
| -               && !setter.IsNull()
 | 
| -               && (setter.usage_counter() > 0)
 | 
| -               && (FLAG_getter_setter_ratio * setter.usage_counter() >=
 | 
| -                   getter.usage_counter());
 | 
| -    if (!result) {
 | 
| -      if (FLAG_trace_optimization) {
 | 
| -        OS::Print("Disabling unboxing of %s\n", field.ToCString());
 | 
| -      }
 | 
| -      field.set_is_unboxing_candidate(false);
 | 
| -      field.DeoptimizeDependentCode();
 | 
| -    } else {
 | 
| -      FlowGraph::AddToGuardedFields(flow_graph_->guarded_fields(), &field);
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::VisitAllocateContext(AllocateContextInstr* instr) {
 | 
| -  // Replace generic allocation with a sequence of inlined allocation and
 | 
| -  // explicit initalizing stores.
 | 
| -  AllocateUninitializedContextInstr* replacement =
 | 
| -      new AllocateUninitializedContextInstr(instr->token_pos(),
 | 
| -                                            instr->num_context_variables());
 | 
| -  instr->ReplaceWith(replacement, current_iterator());
 | 
| -
 | 
| -  StoreInstanceFieldInstr* store =
 | 
| -      new(I) StoreInstanceFieldInstr(Context::parent_offset(),
 | 
| -                                     new Value(replacement),
 | 
| -                                     new Value(flow_graph_->constant_null()),
 | 
| -                                     kNoStoreBarrier,
 | 
| -                                     instr->token_pos());
 | 
| -  store->set_is_initialization(true);  // Won't be eliminated by DSE.
 | 
| -  flow_graph_->InsertAfter(replacement, store, NULL, FlowGraph::kEffect);
 | 
| -  Definition* cursor = store;
 | 
| -  for (intptr_t i = 0; i < instr->num_context_variables(); ++i) {
 | 
| -    store =
 | 
| -        new(I) StoreInstanceFieldInstr(Context::variable_offset(i),
 | 
| -                                       new Value(replacement),
 | 
| -                                       new Value(flow_graph_->constant_null()),
 | 
| -                                       kNoStoreBarrier,
 | 
| -                                       instr->token_pos());
 | 
| -    store->set_is_initialization(true);  // Won't be eliminated by DSE.
 | 
| -    flow_graph_->InsertAfter(cursor, store, NULL, FlowGraph::kEffect);
 | 
| -    cursor = store;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::VisitLoadCodeUnits(LoadCodeUnitsInstr* instr) {
 | 
| -  // TODO(zerny): Use kUnboxedUint32 once it is fully supported/optimized.
 | 
| -#if defined(TARGET_ARCH_IA32) || defined(TARGET_ARCH_ARM)
 | 
| -  if (!instr->can_pack_into_smi())
 | 
| -    instr->set_representation(kUnboxedMint);
 | 
| -#endif
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool FlowGraphOptimizer::TryInlineInstanceSetter(InstanceCallInstr* instr,
 | 
| -                                                 const ICData& unary_ic_data) {
 | 
| -  ASSERT((unary_ic_data.NumberOfChecks() > 0) &&
 | 
| -      (unary_ic_data.NumArgsTested() == 1));
 | 
| -  if (FLAG_enable_type_checks) {
 | 
| -    // Checked mode setters are inlined like normal methods by conventional
 | 
| -    // inlining.
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  ASSERT(instr->HasICData());
 | 
| -  if (unary_ic_data.NumberOfChecks() == 0) {
 | 
| -    // No type feedback collected.
 | 
| -    return false;
 | 
| -  }
 | 
| -  if (!unary_ic_data.HasOneTarget()) {
 | 
| -    // Polymorphic sites are inlined like normal method calls by conventional
 | 
| -    // inlining.
 | 
| -    return false;
 | 
| -  }
 | 
| -  Function& target = Function::Handle(I);
 | 
| -  intptr_t class_id;
 | 
| -  unary_ic_data.GetOneClassCheckAt(0, &class_id, &target);
 | 
| -  if (target.kind() != RawFunction::kImplicitSetter) {
 | 
| -    // Non-implicit setter are inlined like normal method calls.
 | 
| -    return false;
 | 
| -  }
 | 
| -  // Inline implicit instance setter.
 | 
| -  const String& field_name =
 | 
| -      String::Handle(I, Field::NameFromSetter(instr->function_name()));
 | 
| -  const Field& field =
 | 
| -      Field::ZoneHandle(I, GetField(class_id, field_name));
 | 
| -  ASSERT(!field.IsNull());
 | 
| -
 | 
| -  if (InstanceCallNeedsClassCheck(instr, RawFunction::kImplicitSetter)) {
 | 
| -    AddReceiverCheck(instr);
 | 
| -  }
 | 
| -  StoreBarrierType needs_store_barrier = kEmitStoreBarrier;
 | 
| -  if (ArgIsAlways(kSmiCid, *instr->ic_data(), 1)) {
 | 
| -    InsertBefore(instr,
 | 
| -                 new(I) CheckSmiInstr(
 | 
| -                     new(I) Value(instr->ArgumentAt(1)),
 | 
| -                     instr->deopt_id(),
 | 
| -                     instr->token_pos()),
 | 
| -                 instr->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -    needs_store_barrier = kNoStoreBarrier;
 | 
| -  }
 | 
| -
 | 
| -  if (field.guarded_cid() != kDynamicCid) {
 | 
| -    InsertBefore(instr,
 | 
| -                 new(I) GuardFieldClassInstr(
 | 
| -                     new(I) Value(instr->ArgumentAt(1)),
 | 
| -                      field,
 | 
| -                      instr->deopt_id()),
 | 
| -                 instr->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -  }
 | 
| -
 | 
| -  if (field.needs_length_check()) {
 | 
| -    InsertBefore(instr,
 | 
| -                 new(I) GuardFieldLengthInstr(
 | 
| -                     new(I) Value(instr->ArgumentAt(1)),
 | 
| -                      field,
 | 
| -                      instr->deopt_id()),
 | 
| -                 instr->env(),
 | 
| -                 FlowGraph::kEffect);
 | 
| -  }
 | 
| -
 | 
| -  // Field guard was detached.
 | 
| -  StoreInstanceFieldInstr* store = new(I) StoreInstanceFieldInstr(
 | 
| -      field,
 | 
| -      new(I) Value(instr->ArgumentAt(0)),
 | 
| -      new(I) Value(instr->ArgumentAt(1)),
 | 
| -      needs_store_barrier,
 | 
| -      instr->token_pos());
 | 
| -
 | 
| -  if (store->IsUnboxedStore()) {
 | 
| -    FlowGraph::AddToGuardedFields(flow_graph_->guarded_fields(), &field);
 | 
| -  }
 | 
| -
 | 
| -  // Discard the environment from the original instruction because the store
 | 
| -  // can't deoptimize.
 | 
| -  instr->RemoveEnvironment();
 | 
| -  ReplaceCall(instr, store);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -#if defined(TARGET_ARCH_ARM) || defined(TARGET_ARCH_IA32)
 | 
| -// Smi widening pass is only meaningful on platforms where Smi
 | 
| -// is smaller than 32bit. For now only support it on ARM and ia32.
 | 
| -static bool CanBeWidened(BinarySmiOpInstr* smi_op) {
 | 
| -  return BinaryInt32OpInstr::IsSupported(smi_op->op_kind(),
 | 
| -                                         smi_op->left(),
 | 
| -                                         smi_op->right());
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool BenefitsFromWidening(BinarySmiOpInstr* smi_op) {
 | 
| -  // TODO(vegorov): when shifts with non-constants shift count are supported
 | 
| -  // add them here as we save untagging for the count.
 | 
| -  switch (smi_op->op_kind()) {
 | 
| -    case Token::kMUL:
 | 
| -    case Token::kSHR:
 | 
| -      // For kMUL we save untagging of the argument for kSHR
 | 
| -      // we save tagging of the result.
 | 
| -      return true;
 | 
| -
 | 
| -    default:
 | 
| -      return false;
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::WidenSmiToInt32() {
 | 
| -  GrowableArray<BinarySmiOpInstr*> candidates;
 | 
| -
 | 
| -  // Step 1. Collect all instructions that potentially benefit from widening of
 | 
| -  // their operands (or their result) into int32 range.
 | 
| -  for (BlockIterator block_it = flow_graph_->reverse_postorder_iterator();
 | 
| -       !block_it.Done();
 | 
| -       block_it.Advance()) {
 | 
| -    for (ForwardInstructionIterator instr_it(block_it.Current());
 | 
| -         !instr_it.Done();
 | 
| -         instr_it.Advance()) {
 | 
| -      BinarySmiOpInstr* smi_op = instr_it.Current()->AsBinarySmiOp();
 | 
| -      if ((smi_op != NULL) &&
 | 
| -          BenefitsFromWidening(smi_op) &&
 | 
| -          CanBeWidened(smi_op)) {
 | 
| -        candidates.Add(smi_op);
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (candidates.is_empty()) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  // Step 2. For each block in the graph compute which loop it belongs to.
 | 
| -  // We will use this information later during computation of the widening's
 | 
| -  // gain: we are going to assume that only conversion occuring inside the
 | 
| -  // same loop should be counted against the gain, all other conversions
 | 
| -  // can be hoisted and thus cost nothing compared to the loop cost itself.
 | 
| -  const ZoneGrowableArray<BlockEntryInstr*>& loop_headers =
 | 
| -      flow_graph()->LoopHeaders();
 | 
| -
 | 
| -  GrowableArray<intptr_t> loops(flow_graph_->preorder().length());
 | 
| -  for (intptr_t i = 0; i < flow_graph_->preorder().length(); i++) {
 | 
| -    loops.Add(-1);
 | 
| -  }
 | 
| -
 | 
| -  for (intptr_t loop_id = 0; loop_id < loop_headers.length(); ++loop_id) {
 | 
| -    for (BitVector::Iterator loop_it(loop_headers[loop_id]->loop_info());
 | 
| -         !loop_it.Done();
 | 
| -         loop_it.Advance()) {
 | 
| -      loops[loop_it.Current()] = loop_id;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Step 3. For each candidate transitively collect all other BinarySmiOpInstr
 | 
| -  // and PhiInstr that depend on it and that it depends on and count amount of
 | 
| -  // untagging operations that we save in assumption that this whole graph of
 | 
| -  // values is using kUnboxedInt32 representation instead of kTagged.
 | 
| -  // Convert those graphs that have positive gain to kUnboxedInt32.
 | 
| -
 | 
| -  // BitVector containing SSA indexes of all processed definitions. Used to skip
 | 
| -  // those candidates that belong to dependency graph of another candidate.
 | 
| -  BitVector* processed =
 | 
| -      new(I) BitVector(I, flow_graph_->current_ssa_temp_index());
 | 
| -
 | 
| -  // Worklist used to collect dependency graph.
 | 
| -  DefinitionWorklist worklist(flow_graph_, candidates.length());
 | 
| -  for (intptr_t i = 0; i < candidates.length(); i++) {
 | 
| -    BinarySmiOpInstr* op = candidates[i];
 | 
| -    if (op->WasEliminated() || processed->Contains(op->ssa_temp_index())) {
 | 
| -      continue;
 | 
| -    }
 | 
| -
 | 
| -    if (FLAG_trace_smi_widening) {
 | 
| -      OS::Print("analysing candidate: %s\n", op->ToCString());
 | 
| -    }
 | 
| -    worklist.Clear();
 | 
| -    worklist.Add(op);
 | 
| -
 | 
| -    // Collect dependency graph. Note: more items are added to worklist
 | 
| -    // inside this loop.
 | 
| -    intptr_t gain = 0;
 | 
| -    for (intptr_t j = 0; j < worklist.definitions().length(); j++) {
 | 
| -      Definition* defn = worklist.definitions()[j];
 | 
| -
 | 
| -      if (FLAG_trace_smi_widening) {
 | 
| -        OS::Print("> %s\n", defn->ToCString());
 | 
| -      }
 | 
| -
 | 
| -      if (defn->IsBinarySmiOp() &&
 | 
| -          BenefitsFromWidening(defn->AsBinarySmiOp())) {
 | 
| -        gain++;
 | 
| -        if (FLAG_trace_smi_widening) {
 | 
| -          OS::Print("^ [%" Pd "] (o) %s\n", gain, defn->ToCString());
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      const intptr_t defn_loop = loops[defn->GetBlock()->preorder_number()];
 | 
| -
 | 
| -      // Process all inputs.
 | 
| -      for (intptr_t k = 0; k < defn->InputCount(); k++) {
 | 
| -        Definition* input = defn->InputAt(k)->definition();
 | 
| -        if (input->IsBinarySmiOp() &&
 | 
| -            CanBeWidened(input->AsBinarySmiOp())) {
 | 
| -          worklist.Add(input);
 | 
| -        } else if (input->IsPhi() && (input->Type()->ToCid() == kSmiCid)) {
 | 
| -          worklist.Add(input);
 | 
| -        } else if (input->IsBinaryMintOp()) {
 | 
| -          // Mint operation produces untagged result. We avoid tagging.
 | 
| -          gain++;
 | 
| -          if (FLAG_trace_smi_widening) {
 | 
| -            OS::Print("^ [%" Pd "] (i) %s\n", gain, input->ToCString());
 | 
| -          }
 | 
| -        } else if (defn_loop == loops[input->GetBlock()->preorder_number()] &&
 | 
| -                   (input->Type()->ToCid() == kSmiCid)) {
 | 
| -          // Input comes from the same loop, is known to be smi and requires
 | 
| -          // untagging.
 | 
| -          // TODO(vegorov) this heuristic assumes that values that are not
 | 
| -          // known to be smi have to be checked and this check can be
 | 
| -          // coalesced with untagging. Start coalescing them.
 | 
| -          gain--;
 | 
| -          if (FLAG_trace_smi_widening) {
 | 
| -            OS::Print("v [%" Pd "] (i) %s\n", gain, input->ToCString());
 | 
| -          }
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      // Process all uses.
 | 
| -      for (Value* use = defn->input_use_list();
 | 
| -           use != NULL;
 | 
| -           use = use->next_use()) {
 | 
| -        Instruction* instr = use->instruction();
 | 
| -        Definition* use_defn = instr->AsDefinition();
 | 
| -        if (use_defn == NULL) {
 | 
| -          // We assume that tagging before returning or pushing argument costs
 | 
| -          // very little compared to the cost of the return/call itself.
 | 
| -          if (!instr->IsReturn() && !instr->IsPushArgument()) {
 | 
| -            gain--;
 | 
| -            if (FLAG_trace_smi_widening) {
 | 
| -              OS::Print("v [%" Pd "] (u) %s\n",
 | 
| -                        gain,
 | 
| -                        use->instruction()->ToCString());
 | 
| -            }
 | 
| -          }
 | 
| -          continue;
 | 
| -        } else if (use_defn->IsBinarySmiOp() &&
 | 
| -                   CanBeWidened(use_defn->AsBinarySmiOp())) {
 | 
| -          worklist.Add(use_defn);
 | 
| -        } else if (use_defn->IsPhi() &&
 | 
| -                   use_defn->AsPhi()->Type()->ToCid() == kSmiCid) {
 | 
| -          worklist.Add(use_defn);
 | 
| -        } else if (use_defn->IsBinaryMintOp()) {
 | 
| -          // BinaryMintOp requires untagging of its inputs.
 | 
| -          // Converting kUnboxedInt32 to kUnboxedMint is essentially zero cost
 | 
| -          // sign extension operation.
 | 
| -          gain++;
 | 
| -          if (FLAG_trace_smi_widening) {
 | 
| -            OS::Print("^ [%" Pd "] (u) %s\n",
 | 
| -                      gain,
 | 
| -                      use->instruction()->ToCString());
 | 
| -          }
 | 
| -        } else if (defn_loop == loops[instr->GetBlock()->preorder_number()]) {
 | 
| -          gain--;
 | 
| -          if (FLAG_trace_smi_widening) {
 | 
| -            OS::Print("v [%" Pd "] (u) %s\n",
 | 
| -                      gain,
 | 
| -                      use->instruction()->ToCString());
 | 
| -          }
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    processed->AddAll(worklist.contains_vector());
 | 
| -
 | 
| -    if (FLAG_trace_smi_widening) {
 | 
| -      OS::Print("~ %s gain %" Pd "\n", op->ToCString(), gain);
 | 
| -    }
 | 
| -
 | 
| -    if (gain > 0) {
 | 
| -      // We have positive gain from widening. Convert all BinarySmiOpInstr into
 | 
| -      // BinaryInt32OpInstr and set representation of all phis to kUnboxedInt32.
 | 
| -      for (intptr_t j = 0; j < worklist.definitions().length(); j++) {
 | 
| -        Definition* defn = worklist.definitions()[j];
 | 
| -        ASSERT(defn->IsPhi() || defn->IsBinarySmiOp());
 | 
| -
 | 
| -        if (defn->IsBinarySmiOp()) {
 | 
| -          BinarySmiOpInstr* smi_op = defn->AsBinarySmiOp();
 | 
| -          BinaryInt32OpInstr* int32_op = new(I) BinaryInt32OpInstr(
 | 
| -            smi_op->op_kind(),
 | 
| -            smi_op->left()->CopyWithType(),
 | 
| -            smi_op->right()->CopyWithType(),
 | 
| -            smi_op->DeoptimizationTarget());
 | 
| -
 | 
| -          smi_op->ReplaceWith(int32_op, NULL);
 | 
| -        } else if (defn->IsPhi()) {
 | 
| -          defn->AsPhi()->set_representation(kUnboxedInt32);
 | 
| -          ASSERT(defn->Type()->IsInt());
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -#else
 | 
| -void FlowGraphOptimizer::WidenSmiToInt32() {
 | 
| -  // TODO(vegorov) ideally on 64-bit platforms we would like to narrow smi
 | 
| -  // operations to 32-bit where it saves tagging and untagging and allows
 | 
| -  // to use shorted (and faster) instructions. But we currently don't
 | 
| -  // save enough range information in the ICData to drive this decision.
 | 
| -}
 | 
| -#endif
 | 
| -
 | 
| -void FlowGraphOptimizer::InferIntRanges() {
 | 
| -  RangeAnalysis range_analysis(flow_graph_);
 | 
| -  range_analysis.Analyze();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void TryCatchAnalyzer::Optimize(FlowGraph* flow_graph) {
 | 
| -  // For every catch-block: Iterate over all call instructions inside the
 | 
| -  // corresponding try-block and figure out for each environment value if it
 | 
| -  // is the same constant at all calls. If yes, replace the initial definition
 | 
| -  // at the catch-entry with this constant.
 | 
| -  const GrowableArray<CatchBlockEntryInstr*>& catch_entries =
 | 
| -      flow_graph->graph_entry()->catch_entries();
 | 
| -  intptr_t base = kFirstLocalSlotFromFp + flow_graph->num_non_copied_params();
 | 
| -  for (intptr_t catch_idx = 0;
 | 
| -       catch_idx < catch_entries.length();
 | 
| -       ++catch_idx) {
 | 
| -    CatchBlockEntryInstr* catch_entry = catch_entries[catch_idx];
 | 
| -
 | 
| -    // Initialize cdefs with the original initial definitions (ParameterInstr).
 | 
| -    // The following representation is used:
 | 
| -    // ParameterInstr => unknown
 | 
| -    // ConstantInstr => known constant
 | 
| -    // NULL => non-constant
 | 
| -    GrowableArray<Definition*>* idefs = catch_entry->initial_definitions();
 | 
| -    GrowableArray<Definition*> cdefs(idefs->length());
 | 
| -    cdefs.AddArray(*idefs);
 | 
| -
 | 
| -    // exception_var and stacktrace_var are never constant.
 | 
| -    intptr_t ex_idx = base - catch_entry->exception_var().index();
 | 
| -    intptr_t st_idx = base - catch_entry->stacktrace_var().index();
 | 
| -    cdefs[ex_idx] = cdefs[st_idx] = NULL;
 | 
| -
 | 
| -    for (BlockIterator block_it = flow_graph->reverse_postorder_iterator();
 | 
| -         !block_it.Done();
 | 
| -         block_it.Advance()) {
 | 
| -      BlockEntryInstr* block = block_it.Current();
 | 
| -      if (block->try_index() == catch_entry->catch_try_index()) {
 | 
| -        for (ForwardInstructionIterator instr_it(block);
 | 
| -             !instr_it.Done();
 | 
| -             instr_it.Advance()) {
 | 
| -          Instruction* current = instr_it.Current();
 | 
| -          if (current->MayThrow()) {
 | 
| -            Environment* env = current->env()->Outermost();
 | 
| -            ASSERT(env != NULL);
 | 
| -            for (intptr_t env_idx = 0; env_idx < cdefs.length(); ++env_idx) {
 | 
| -              if (cdefs[env_idx] != NULL &&
 | 
| -                  env->ValueAt(env_idx)->BindsToConstant()) {
 | 
| -                cdefs[env_idx] = env->ValueAt(env_idx)->definition();
 | 
| -              }
 | 
| -              if (cdefs[env_idx] != env->ValueAt(env_idx)->definition()) {
 | 
| -                cdefs[env_idx] = NULL;
 | 
| -              }
 | 
| -            }
 | 
| -          }
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -    for (intptr_t j = 0; j < idefs->length(); ++j) {
 | 
| -      if (cdefs[j] != NULL && cdefs[j]->IsConstant()) {
 | 
| -        // TODO(fschneider): Use constants from the constant pool.
 | 
| -        Definition* old = (*idefs)[j];
 | 
| -        ConstantInstr* orig = cdefs[j]->AsConstant();
 | 
| -        ConstantInstr* copy =
 | 
| -            new(flow_graph->isolate()) ConstantInstr(orig->value());
 | 
| -        copy->set_ssa_temp_index(flow_graph->alloc_ssa_temp_index());
 | 
| -        old->ReplaceUsesWith(copy);
 | 
| -        (*idefs)[j] = copy;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -LICM::LICM(FlowGraph* flow_graph) : flow_graph_(flow_graph) {
 | 
| -  ASSERT(flow_graph->is_licm_allowed());
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void LICM::Hoist(ForwardInstructionIterator* it,
 | 
| -                 BlockEntryInstr* pre_header,
 | 
| -                 Instruction* current) {
 | 
| -  if (current->IsCheckClass()) {
 | 
| -    current->AsCheckClass()->set_licm_hoisted(true);
 | 
| -  } else if (current->IsCheckSmi()) {
 | 
| -    current->AsCheckSmi()->set_licm_hoisted(true);
 | 
| -  } else if (current->IsCheckEitherNonSmi()) {
 | 
| -    current->AsCheckEitherNonSmi()->set_licm_hoisted(true);
 | 
| -  } else if (current->IsCheckArrayBound()) {
 | 
| -    current->AsCheckArrayBound()->set_licm_hoisted(true);
 | 
| -  }
 | 
| -  if (FLAG_trace_optimization) {
 | 
| -    OS::Print("Hoisting instruction %s:%" Pd " from B%" Pd " to B%" Pd "\n",
 | 
| -              current->DebugName(),
 | 
| -              current->GetDeoptId(),
 | 
| -              current->GetBlock()->block_id(),
 | 
| -              pre_header->block_id());
 | 
| -  }
 | 
| -  // Move the instruction out of the loop.
 | 
| -  current->RemoveEnvironment();
 | 
| -  if (it != NULL) {
 | 
| -    it->RemoveCurrentFromGraph();
 | 
| -  } else {
 | 
| -    current->RemoveFromGraph();
 | 
| -  }
 | 
| -  GotoInstr* last = pre_header->last_instruction()->AsGoto();
 | 
| -  // Using kind kEffect will not assign a fresh ssa temporary index.
 | 
| -  flow_graph()->InsertBefore(last, current, last->env(), FlowGraph::kEffect);
 | 
| -  current->CopyDeoptIdFrom(*last);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void LICM::TrySpecializeSmiPhi(PhiInstr* phi,
 | 
| -                               BlockEntryInstr* header,
 | 
| -                               BlockEntryInstr* pre_header) {
 | 
| -  if (phi->Type()->ToCid() == kSmiCid) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  // Check if there is only a single kDynamicCid input to the phi that
 | 
| -  // comes from the pre-header.
 | 
| -  const intptr_t kNotFound = -1;
 | 
| -  intptr_t non_smi_input = kNotFound;
 | 
| -  for (intptr_t i = 0; i < phi->InputCount(); ++i) {
 | 
| -    Value* input = phi->InputAt(i);
 | 
| -    if (input->Type()->ToCid() != kSmiCid) {
 | 
| -      if ((non_smi_input != kNotFound) ||
 | 
| -          (input->Type()->ToCid() != kDynamicCid)) {
 | 
| -        // There are multiple kDynamicCid inputs or there is an input that is
 | 
| -        // known to be non-smi.
 | 
| -        return;
 | 
| -      } else {
 | 
| -        non_smi_input = i;
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if ((non_smi_input == kNotFound) ||
 | 
| -      (phi->block()->PredecessorAt(non_smi_input) != pre_header)) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  CheckSmiInstr* check = NULL;
 | 
| -  for (Value* use = phi->input_use_list();
 | 
| -       (use != NULL) && (check == NULL);
 | 
| -       use = use->next_use()) {
 | 
| -    check = use->instruction()->AsCheckSmi();
 | 
| -  }
 | 
| -
 | 
| -  if (check == NULL) {
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  // Host CheckSmi instruction and make this phi smi one.
 | 
| -  Hoist(NULL, pre_header, check);
 | 
| -
 | 
| -  // Replace value we are checking with phi's input.
 | 
| -  check->value()->BindTo(phi->InputAt(non_smi_input)->definition());
 | 
| -
 | 
| -  phi->UpdateType(CompileType::FromCid(kSmiCid));
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Load instructions handled by load elimination.
 | 
| -static bool IsLoadEliminationCandidate(Instruction* instr) {
 | 
| -  return instr->IsLoadField()
 | 
| -      || instr->IsLoadIndexed()
 | 
| -      || instr->IsLoadStaticField();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool IsLoopInvariantLoad(ZoneGrowableArray<BitVector*>* sets,
 | 
| -                                intptr_t loop_header_index,
 | 
| -                                Instruction* instr) {
 | 
| -  return IsLoadEliminationCandidate(instr) &&
 | 
| -      (sets != NULL) &&
 | 
| -      instr->HasPlaceId() &&
 | 
| -      ((*sets)[loop_header_index] != NULL) &&
 | 
| -      (*sets)[loop_header_index]->Contains(instr->place_id());
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void LICM::OptimisticallySpecializeSmiPhis() {
 | 
| -  if (!flow_graph()->parsed_function()->function().
 | 
| -          allows_hoisting_check_class()) {
 | 
| -    // Do not hoist any.
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  const ZoneGrowableArray<BlockEntryInstr*>& loop_headers =
 | 
| -      flow_graph()->LoopHeaders();
 | 
| -
 | 
| -  for (intptr_t i = 0; i < loop_headers.length(); ++i) {
 | 
| -    JoinEntryInstr* header = loop_headers[i]->AsJoinEntry();
 | 
| -    // Skip loop that don't have a pre-header block.
 | 
| -    BlockEntryInstr* pre_header = header->ImmediateDominator();
 | 
| -    if (pre_header == NULL) continue;
 | 
| -
 | 
| -    for (PhiIterator it(header); !it.Done(); it.Advance()) {
 | 
| -      TrySpecializeSmiPhi(it.Current(), header, pre_header);
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void LICM::Optimize() {
 | 
| -  if (!flow_graph()->parsed_function()->function().
 | 
| -          allows_hoisting_check_class()) {
 | 
| -    // Do not hoist any.
 | 
| -    return;
 | 
| -  }
 | 
| -
 | 
| -  const ZoneGrowableArray<BlockEntryInstr*>& loop_headers =
 | 
| -      flow_graph()->LoopHeaders();
 | 
| -
 | 
| -  ZoneGrowableArray<BitVector*>* loop_invariant_loads =
 | 
| -      flow_graph()->loop_invariant_loads();
 | 
| -
 | 
| -  BlockEffects* block_effects = flow_graph()->block_effects();
 | 
| -
 | 
| -  for (intptr_t i = 0; i < loop_headers.length(); ++i) {
 | 
| -    BlockEntryInstr* header = loop_headers[i];
 | 
| -    // Skip loop that don't have a pre-header block.
 | 
| -    BlockEntryInstr* pre_header = header->ImmediateDominator();
 | 
| -    if (pre_header == NULL) continue;
 | 
| -
 | 
| -    for (BitVector::Iterator loop_it(header->loop_info());
 | 
| -         !loop_it.Done();
 | 
| -         loop_it.Advance()) {
 | 
| -      BlockEntryInstr* block = flow_graph()->preorder()[loop_it.Current()];
 | 
| -      for (ForwardInstructionIterator it(block);
 | 
| -           !it.Done();
 | 
| -           it.Advance()) {
 | 
| -        Instruction* current = it.Current();
 | 
| -        if ((current->AllowsCSE() &&
 | 
| -             block_effects->CanBeMovedTo(current, pre_header)) ||
 | 
| -            IsLoopInvariantLoad(loop_invariant_loads, i, current)) {
 | 
| -          bool inputs_loop_invariant = true;
 | 
| -          for (int i = 0; i < current->InputCount(); ++i) {
 | 
| -            Definition* input_def = current->InputAt(i)->definition();
 | 
| -            if (!input_def->GetBlock()->Dominates(pre_header)) {
 | 
| -              inputs_loop_invariant = false;
 | 
| -              break;
 | 
| -            }
 | 
| -          }
 | 
| -          if (inputs_loop_invariant &&
 | 
| -              !current->IsAssertAssignable() &&
 | 
| -              !current->IsAssertBoolean()) {
 | 
| -            // TODO(fschneider): Enable hoisting of Assert-instructions
 | 
| -            // if it safe to do.
 | 
| -            Hoist(&it, pre_header, current);
 | 
| -          }
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Place describes an abstract location (e.g. field) that IR can load
 | 
| -// from or store to.
 | 
| -//
 | 
| -// Places are also used to describe wild-card locations also known as aliases,
 | 
| -// that essentially represent sets of places that alias each other. Places A
 | 
| -// and B are said to alias each other if store into A can affect load from B.
 | 
| -//
 | 
| -// We distinguish the following aliases:
 | 
| -//
 | 
| -//   - for fields
 | 
| -//     - *.f, *.@offs - field inside some object;
 | 
| -//     - X.f, X.@offs - field inside an allocated object X;
 | 
| -//   - for indexed accesses
 | 
| -//     - *[*] - non-constant index inside some object;
 | 
| -//     - *[C] - constant index inside some object;
 | 
| -//     - X[*] - non-constant index inside an allocated object X;
 | 
| -//     - X[C] - constant index inside an allocated object X.
 | 
| -//
 | 
| -// Separating allocations from other objects improves precision of the
 | 
| -// load forwarding pass because of the following two properties:
 | 
| -//
 | 
| -//   - if X can be proven to have no aliases itself (i.e. there is no other SSA
 | 
| -//     variable that points to X) then no place inside X can be aliased with any
 | 
| -//     wildcard dependent place (*.f, *.@offs, *[*], *[C]);
 | 
| -//   - given allocations X and Y no place inside X can be aliased with any place
 | 
| -//     inside Y even if any of them or both escape.
 | 
| -//
 | 
| -// It important to realize that single place can belong to multiple aliases.
 | 
| -// For example place X.f with aliased allocation X belongs both to X.f and *.f
 | 
| -// aliases. Likewise X[C] with non-aliased allocation X belongs to X[C] and X[*]
 | 
| -// aliases.
 | 
| -//
 | 
| -class Place : public ValueObject {
 | 
| - public:
 | 
| -  enum Kind {
 | 
| -    kNone,
 | 
| -
 | 
| -    // Field location. For instance fields is represented as a pair of a Field
 | 
| -    // object and an instance (SSA definition) that is being accessed.
 | 
| -    // For static fields instance is NULL.
 | 
| -    kField,
 | 
| -
 | 
| -    // VMField location. Represented as a pair of an instance (SSA definition)
 | 
| -    // being accessed and offset to the field.
 | 
| -    kVMField,
 | 
| -
 | 
| -    // Indexed location with a non-constant index.
 | 
| -    kIndexed,
 | 
| -
 | 
| -    // Indexed location with a constant index.
 | 
| -    kConstantIndexed,
 | 
| -  };
 | 
| -
 | 
| -  Place(const Place& other)
 | 
| -      : ValueObject(),
 | 
| -        kind_(other.kind_),
 | 
| -        representation_(other.representation_),
 | 
| -        instance_(other.instance_),
 | 
| -        raw_selector_(other.raw_selector_),
 | 
| -        id_(other.id_) {
 | 
| -  }
 | 
| -
 | 
| -  // Construct a place from instruction if instruction accesses any place.
 | 
| -  // Otherwise constructs kNone place.
 | 
| -  Place(Instruction* instr, bool* is_load, bool* is_store)
 | 
| -      : kind_(kNone),
 | 
| -        representation_(kNoRepresentation),
 | 
| -        instance_(NULL),
 | 
| -        raw_selector_(0),
 | 
| -        id_(0) {
 | 
| -    switch (instr->tag()) {
 | 
| -      case Instruction::kLoadField: {
 | 
| -        LoadFieldInstr* load_field = instr->AsLoadField();
 | 
| -        representation_ = load_field->representation();
 | 
| -        instance_ = load_field->instance()->definition()->OriginalDefinition();
 | 
| -        if (load_field->field() != NULL) {
 | 
| -          kind_ = kField;
 | 
| -          field_ = load_field->field();
 | 
| -        } else {
 | 
| -          kind_ = kVMField;
 | 
| -          offset_in_bytes_ = load_field->offset_in_bytes();
 | 
| -        }
 | 
| -        *is_load = true;
 | 
| -        break;
 | 
| -      }
 | 
| -
 | 
| -      case Instruction::kStoreInstanceField: {
 | 
| -        StoreInstanceFieldInstr* store =
 | 
| -            instr->AsStoreInstanceField();
 | 
| -        representation_ = store->RequiredInputRepresentation(
 | 
| -            StoreInstanceFieldInstr::kValuePos);
 | 
| -        instance_ = store->instance()->definition()->OriginalDefinition();
 | 
| -        if (!store->field().IsNull()) {
 | 
| -          kind_ = kField;
 | 
| -          field_ = &store->field();
 | 
| -        } else {
 | 
| -          kind_ = kVMField;
 | 
| -          offset_in_bytes_ = store->offset_in_bytes();
 | 
| -        }
 | 
| -        *is_store = true;
 | 
| -        break;
 | 
| -      }
 | 
| -
 | 
| -      case Instruction::kLoadStaticField:
 | 
| -        kind_ = kField;
 | 
| -        representation_ = instr->AsLoadStaticField()->representation();
 | 
| -        field_ = &instr->AsLoadStaticField()->StaticField();
 | 
| -        *is_load = true;
 | 
| -        break;
 | 
| -
 | 
| -      case Instruction::kStoreStaticField:
 | 
| -        kind_ = kField;
 | 
| -        representation_ = instr->AsStoreStaticField()->
 | 
| -            RequiredInputRepresentation(StoreStaticFieldInstr::kValuePos);
 | 
| -        field_ = &instr->AsStoreStaticField()->field();
 | 
| -        *is_store = true;
 | 
| -        break;
 | 
| -
 | 
| -      case Instruction::kLoadIndexed: {
 | 
| -        LoadIndexedInstr* load_indexed = instr->AsLoadIndexed();
 | 
| -        representation_ = load_indexed->representation();
 | 
| -        instance_ = load_indexed->array()->definition()->OriginalDefinition();
 | 
| -        SetIndex(load_indexed->index()->definition());
 | 
| -        *is_load = true;
 | 
| -        break;
 | 
| -      }
 | 
| -
 | 
| -      case Instruction::kStoreIndexed: {
 | 
| -        StoreIndexedInstr* store_indexed = instr->AsStoreIndexed();
 | 
| -        representation_ = store_indexed->
 | 
| -            RequiredInputRepresentation(StoreIndexedInstr::kValuePos);
 | 
| -        instance_ = store_indexed->array()->definition()->OriginalDefinition();
 | 
| -        SetIndex(store_indexed->index()->definition());
 | 
| -        *is_store = true;
 | 
| -        break;
 | 
| -      }
 | 
| -
 | 
| -      default:
 | 
| -        break;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Create object representing *[*] alias.
 | 
| -  static Place* CreateAnyInstanceAnyIndexAlias(Isolate* isolate,
 | 
| -                                               intptr_t id) {
 | 
| -    return Wrap(isolate, Place(kIndexed, NULL, 0), id);
 | 
| -  }
 | 
| -
 | 
| -  // Return least generic alias for this place. Given that aliases are
 | 
| -  // essentially sets of places we define least generic alias as a smallest
 | 
| -  // alias that contains this place.
 | 
| -  //
 | 
| -  // We obtain such alias by a simple transformation:
 | 
| -  //
 | 
| -  //    - for places that depend on an instance X.f, X.@offs, X[i], X[C]
 | 
| -  //      we drop X if X is not an allocation because in this case X does not
 | 
| -  //      posess an identity obtaining aliases *.f, *.@offs, *[i] and *[C]
 | 
| -  //      respectively;
 | 
| -  //    - for non-constant indexed places X[i] we drop information about the
 | 
| -  //      index obtaining alias X[*].
 | 
| -  //
 | 
| -  Place ToAlias() const {
 | 
| -    return Place(
 | 
| -        kind_,
 | 
| -        (DependsOnInstance() && IsAllocation(instance())) ? instance() : NULL,
 | 
| -        (kind() == kIndexed) ? 0 : raw_selector_);
 | 
| -  }
 | 
| -
 | 
| -  bool DependsOnInstance() const {
 | 
| -    switch (kind()) {
 | 
| -      case kField:
 | 
| -      case kVMField:
 | 
| -      case kIndexed:
 | 
| -      case kConstantIndexed:
 | 
| -        return true;
 | 
| -
 | 
| -      case kNone:
 | 
| -        return false;
 | 
| -    }
 | 
| -
 | 
| -    UNREACHABLE();
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  // Given instance dependent alias X.f, X.@offs, X[C], X[*] return
 | 
| -  // wild-card dependent alias *.f, *.@offs, *[C] or *[*] respectively.
 | 
| -  Place CopyWithoutInstance() const {
 | 
| -    ASSERT(DependsOnInstance());
 | 
| -    return Place(kind_, NULL, raw_selector_);
 | 
| -  }
 | 
| -
 | 
| -  // Given alias X[C] or *[C] return X[*] and *[*] respectively.
 | 
| -  Place CopyWithoutIndex() const {
 | 
| -    ASSERT(kind_ == kConstantIndexed);
 | 
| -    return Place(kIndexed, instance_, 0);
 | 
| -  }
 | 
| -
 | 
| -  intptr_t id() const { return id_; }
 | 
| -
 | 
| -  Kind kind() const { return kind_; }
 | 
| -
 | 
| -  Representation representation() const { return representation_; }
 | 
| -
 | 
| -  Definition* instance() const {
 | 
| -    ASSERT(DependsOnInstance());
 | 
| -    return instance_;
 | 
| -  }
 | 
| -
 | 
| -  void set_instance(Definition* def) {
 | 
| -    ASSERT(DependsOnInstance());
 | 
| -    instance_ = def->OriginalDefinition();
 | 
| -  }
 | 
| -
 | 
| -  const Field& field() const {
 | 
| -    ASSERT(kind_ == kField);
 | 
| -    return *field_;
 | 
| -  }
 | 
| -
 | 
| -  intptr_t offset_in_bytes() const {
 | 
| -    ASSERT(kind_ == kVMField);
 | 
| -    return offset_in_bytes_;
 | 
| -  }
 | 
| -
 | 
| -  Definition* index() const {
 | 
| -    ASSERT(kind_ == kIndexed);
 | 
| -    return index_;
 | 
| -  }
 | 
| -
 | 
| -  intptr_t index_constant() const {
 | 
| -    ASSERT(kind_ == kConstantIndexed);
 | 
| -    return index_constant_;
 | 
| -  }
 | 
| -
 | 
| -  static const char* DefinitionName(Definition* def) {
 | 
| -    if (def == NULL) {
 | 
| -      return "*";
 | 
| -    } else {
 | 
| -      return Isolate::Current()->current_zone()->PrintToString(
 | 
| -            "v%" Pd, def->ssa_temp_index());
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  const char* ToCString() const {
 | 
| -    switch (kind_) {
 | 
| -      case kNone:
 | 
| -        return "<none>";
 | 
| -
 | 
| -      case kField: {
 | 
| -        const char* field_name = String::Handle(field().name()).ToCString();
 | 
| -        if (field().is_static()) {
 | 
| -          return Isolate::Current()->current_zone()->PrintToString(
 | 
| -              "<%s>", field_name);
 | 
| -        } else {
 | 
| -          return Isolate::Current()->current_zone()->PrintToString(
 | 
| -              "<%s.%s>", DefinitionName(instance()), field_name);
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      case kVMField:
 | 
| -        return Isolate::Current()->current_zone()->PrintToString(
 | 
| -            "<%s.@%" Pd ">",
 | 
| -            DefinitionName(instance()),
 | 
| -            offset_in_bytes());
 | 
| -
 | 
| -      case kIndexed:
 | 
| -        return Isolate::Current()->current_zone()->PrintToString(
 | 
| -            "<%s[%s]>",
 | 
| -            DefinitionName(instance()),
 | 
| -            DefinitionName(index()));
 | 
| -
 | 
| -      case kConstantIndexed:
 | 
| -        return Isolate::Current()->current_zone()->PrintToString(
 | 
| -            "<%s[%" Pd "]>",
 | 
| -            DefinitionName(instance()),
 | 
| -            index_constant());
 | 
| -    }
 | 
| -    UNREACHABLE();
 | 
| -    return "<?>";
 | 
| -  }
 | 
| -
 | 
| -  bool IsFinalField() const {
 | 
| -    return (kind() == kField) && field().is_final();
 | 
| -  }
 | 
| -
 | 
| -  intptr_t Hashcode() const {
 | 
| -    return (kind_ * 63 + reinterpret_cast<intptr_t>(instance_)) * 31 +
 | 
| -        representation_ * 15 + FieldHashcode();
 | 
| -  }
 | 
| -
 | 
| -  bool Equals(const Place* other) const {
 | 
| -    return (kind_ == other->kind_) &&
 | 
| -        (representation_ == other->representation_) &&
 | 
| -        (instance_ == other->instance_) &&
 | 
| -        SameField(other);
 | 
| -  }
 | 
| -
 | 
| -  // Create a zone allocated copy of this place and assign given id to it.
 | 
| -  static Place* Wrap(Isolate* isolate, const Place& place, intptr_t id);
 | 
| -
 | 
| -  static bool IsAllocation(Definition* defn) {
 | 
| -    return (defn != NULL) &&
 | 
| -        (defn->IsAllocateObject() ||
 | 
| -         defn->IsCreateArray() ||
 | 
| -         defn->IsAllocateUninitializedContext() ||
 | 
| -         (defn->IsStaticCall() &&
 | 
| -          defn->AsStaticCall()->IsRecognizedFactory()));
 | 
| -  }
 | 
| -
 | 
| - private:
 | 
| -  Place(Kind kind, Definition* instance, intptr_t selector)
 | 
| -      : kind_(kind),
 | 
| -        representation_(kNoRepresentation),
 | 
| -        instance_(instance),
 | 
| -        raw_selector_(selector),
 | 
| -        id_(0) {
 | 
| -  }
 | 
| -
 | 
| -  bool SameField(const Place* other) const {
 | 
| -    return (kind_ == kField) ? (field().raw() == other->field().raw())
 | 
| -                             : (offset_in_bytes_ == other->offset_in_bytes_);
 | 
| -  }
 | 
| -
 | 
| -  intptr_t FieldHashcode() const {
 | 
| -    return (kind_ == kField) ? reinterpret_cast<intptr_t>(field().raw())
 | 
| -                             : offset_in_bytes_;
 | 
| -  }
 | 
| -
 | 
| -  void SetIndex(Definition* index) {
 | 
| -    ConstantInstr* index_constant = index->AsConstant();
 | 
| -    if ((index_constant != NULL) && index_constant->value().IsSmi()) {
 | 
| -      kind_ = kConstantIndexed;
 | 
| -      index_constant_ = Smi::Cast(index_constant->value()).Value();
 | 
| -    } else {
 | 
| -      kind_ = kIndexed;
 | 
| -      index_ = index;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  Kind kind_;
 | 
| -  Representation representation_;
 | 
| -  Definition* instance_;
 | 
| -  union {
 | 
| -    intptr_t raw_selector_;
 | 
| -    const Field* field_;
 | 
| -    intptr_t offset_in_bytes_;
 | 
| -    intptr_t index_constant_;
 | 
| -    Definition* index_;
 | 
| -  };
 | 
| -
 | 
| -  intptr_t id_;
 | 
| -};
 | 
| -
 | 
| -
 | 
| -class ZonePlace : public ZoneAllocated {
 | 
| - public:
 | 
| -  explicit ZonePlace(const Place& place) : place_(place) { }
 | 
| -
 | 
| -  Place* place() { return &place_; }
 | 
| -
 | 
| - private:
 | 
| -  Place place_;
 | 
| -};
 | 
| -
 | 
| -
 | 
| -Place* Place::Wrap(Isolate* isolate, const Place& place, intptr_t id) {
 | 
| -  Place* wrapped = (new(isolate) ZonePlace(place))->place();
 | 
| -  wrapped->id_ = id;
 | 
| -  return wrapped;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Correspondence between places connected through outgoing phi moves on the
 | 
| -// edge that targets join.
 | 
| -class PhiPlaceMoves : public ZoneAllocated {
 | 
| - public:
 | 
| -  // Record a move from the place with id |from| to the place with id |to| at
 | 
| -  // the given block.
 | 
| -  void CreateOutgoingMove(Isolate* isolate,
 | 
| -                          BlockEntryInstr* block, intptr_t from, intptr_t to) {
 | 
| -    const intptr_t block_num = block->preorder_number();
 | 
| -    while (moves_.length() <= block_num) {
 | 
| -      moves_.Add(NULL);
 | 
| -    }
 | 
| -
 | 
| -    if (moves_[block_num] == NULL) {
 | 
| -      moves_[block_num] = new(isolate) ZoneGrowableArray<Move>(5);
 | 
| -    }
 | 
| -
 | 
| -    moves_[block_num]->Add(Move(from, to));
 | 
| -  }
 | 
| -
 | 
| -  class Move {
 | 
| -   public:
 | 
| -    Move(intptr_t from, intptr_t to) : from_(from), to_(to) { }
 | 
| -
 | 
| -    intptr_t from() const { return from_; }
 | 
| -    intptr_t to() const { return to_; }
 | 
| -
 | 
| -   private:
 | 
| -    intptr_t from_;
 | 
| -    intptr_t to_;
 | 
| -  };
 | 
| -
 | 
| -  typedef const ZoneGrowableArray<Move>* MovesList;
 | 
| -
 | 
| -  MovesList GetOutgoingMoves(BlockEntryInstr* block) const {
 | 
| -    const intptr_t block_num = block->preorder_number();
 | 
| -    return (block_num < moves_.length()) ?
 | 
| -        moves_[block_num] : NULL;
 | 
| -  }
 | 
| -
 | 
| - private:
 | 
| -  GrowableArray<ZoneGrowableArray<Move>* > moves_;
 | 
| -};
 | 
| -
 | 
| -
 | 
| -// A map from aliases to a set of places sharing the alias. Additionally
 | 
| -// carries a set of places that can be aliased by side-effects, essentially
 | 
| -// those that are affected by calls.
 | 
| -class AliasedSet : public ZoneAllocated {
 | 
| - public:
 | 
| -  AliasedSet(Isolate* isolate,
 | 
| -             DirectChainedHashMap<PointerKeyValueTrait<Place> >* places_map,
 | 
| -             ZoneGrowableArray<Place*>* places,
 | 
| -             PhiPlaceMoves* phi_moves)
 | 
| -      : isolate_(isolate),
 | 
| -        places_map_(places_map),
 | 
| -        places_(*places),
 | 
| -        phi_moves_(phi_moves),
 | 
| -        aliases_(5),
 | 
| -        aliases_map_(),
 | 
| -        representatives_(),
 | 
| -        killed_(),
 | 
| -        aliased_by_effects_(new(isolate) BitVector(isolate, places->length())) {
 | 
| -    InsertAlias(Place::CreateAnyInstanceAnyIndexAlias(isolate_,
 | 
| -        kAnyInstanceAnyIndexAlias));
 | 
| -    for (intptr_t i = 0; i < places_.length(); i++) {
 | 
| -      AddRepresentative(places_[i]);
 | 
| -    }
 | 
| -    ComputeKillSets();
 | 
| -  }
 | 
| -
 | 
| -  intptr_t LookupAliasId(const Place& alias) {
 | 
| -    const Place* result = aliases_map_.Lookup(&alias);
 | 
| -    return (result != NULL) ? result->id() : static_cast<intptr_t>(kNoAlias);
 | 
| -  }
 | 
| -
 | 
| -  BitVector* GetKilledSet(intptr_t alias) {
 | 
| -    return (alias < killed_.length()) ? killed_[alias] : NULL;
 | 
| -  }
 | 
| -
 | 
| -  intptr_t max_place_id() const { return places().length(); }
 | 
| -  bool IsEmpty() const { return max_place_id() == 0; }
 | 
| -
 | 
| -  BitVector* aliased_by_effects() const { return aliased_by_effects_; }
 | 
| -
 | 
| -  const ZoneGrowableArray<Place*>& places() const {
 | 
| -    return places_;
 | 
| -  }
 | 
| -
 | 
| -  Place* LookupCanonical(Place* place) const {
 | 
| -    return places_map_->Lookup(place);
 | 
| -  }
 | 
| -
 | 
| -  void PrintSet(BitVector* set) {
 | 
| -    bool comma = false;
 | 
| -    for (BitVector::Iterator it(set);
 | 
| -         !it.Done();
 | 
| -         it.Advance()) {
 | 
| -      if (comma) {
 | 
| -        OS::Print(", ");
 | 
| -      }
 | 
| -      OS::Print("%s", places_[it.Current()]->ToCString());
 | 
| -      comma = true;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  const PhiPlaceMoves* phi_moves() const { return phi_moves_; }
 | 
| -
 | 
| -  void RollbackAliasedIdentites() {
 | 
| -    for (intptr_t i = 0; i < identity_rollback_.length(); ++i) {
 | 
| -      identity_rollback_[i]->SetIdentity(AliasIdentity::Unknown());
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Returns false if the result of an allocation instruction can't be aliased
 | 
| -  // by another SSA variable and true otherwise.
 | 
| -  bool CanBeAliased(Definition* alloc) {
 | 
| -    if (!Place::IsAllocation(alloc)) {
 | 
| -      return true;
 | 
| -    }
 | 
| -
 | 
| -    if (alloc->Identity().IsUnknown()) {
 | 
| -      ComputeAliasing(alloc);
 | 
| -    }
 | 
| -
 | 
| -    return !alloc->Identity().IsNotAliased();
 | 
| -  }
 | 
| -
 | 
| -  enum {
 | 
| -    kNoAlias = 0
 | 
| -  };
 | 
| -
 | 
| - private:
 | 
| -  enum {
 | 
| -    // Artificial alias that is used to collect all representatives of the
 | 
| -    // *[C], X[C] aliases for arbitrary C.
 | 
| -    kAnyConstantIndexedAlias = 1,
 | 
| -
 | 
| -    // Artificial alias that is used to collect all representatives of
 | 
| -    // *[C] alias for arbitrary C.
 | 
| -    kUnknownInstanceConstantIndexedAlias = 2,
 | 
| -
 | 
| -    // Artificial alias that is used to collect all representatives of
 | 
| -    // X[*] alias for all X.
 | 
| -    kAnyAllocationIndexedAlias = 3,
 | 
| -
 | 
| -    // *[*] alias.
 | 
| -    kAnyInstanceAnyIndexAlias = 4
 | 
| -  };
 | 
| -
 | 
| -  // Compute least generic alias for the place and assign alias id to it.
 | 
| -  void AddRepresentative(Place* place) {
 | 
| -    if (!place->IsFinalField()) {
 | 
| -      const Place* alias = CanonicalizeAlias(place->ToAlias());
 | 
| -      EnsureSet(&representatives_, alias->id())->Add(place->id());
 | 
| -
 | 
| -      // Update cumulative representative sets that are used during
 | 
| -      // killed sets computation.
 | 
| -      if (alias->kind() == Place::kConstantIndexed) {
 | 
| -        if (CanBeAliased(alias->instance())) {
 | 
| -          EnsureSet(&representatives_, kAnyConstantIndexedAlias)->
 | 
| -              Add(place->id());
 | 
| -        }
 | 
| -
 | 
| -        if (alias->instance() == NULL) {
 | 
| -          EnsureSet(&representatives_, kUnknownInstanceConstantIndexedAlias)->
 | 
| -              Add(place->id());
 | 
| -        }
 | 
| -      } else if ((alias->kind() == Place::kIndexed) &&
 | 
| -                 CanBeAliased(place->instance())) {
 | 
| -        EnsureSet(&representatives_, kAnyAllocationIndexedAlias)->
 | 
| -            Add(place->id());
 | 
| -      }
 | 
| -
 | 
| -      if (!IsIndependentFromEffects(place)) {
 | 
| -        aliased_by_effects_->Add(place->id());
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  void ComputeKillSets() {
 | 
| -    for (intptr_t i = 0; i < aliases_.length(); ++i) {
 | 
| -      const Place* alias = aliases_[i];
 | 
| -      // Add all representatives to the kill set.
 | 
| -      AddAllRepresentatives(alias->id(), alias->id());
 | 
| -      ComputeKillSet(alias);
 | 
| -    }
 | 
| -
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      OS::Print("Aliases KILL sets:\n");
 | 
| -      for (intptr_t i = 0; i < aliases_.length(); ++i) {
 | 
| -        const Place* alias = aliases_[i];
 | 
| -        BitVector* kill = GetKilledSet(alias->id());
 | 
| -
 | 
| -        OS::Print("%s: ", alias->ToCString());
 | 
| -        if (kill != NULL) {
 | 
| -          PrintSet(kill);
 | 
| -        }
 | 
| -        OS::Print("\n");
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  void InsertAlias(const Place* alias) {
 | 
| -    aliases_map_.Insert(alias);
 | 
| -    aliases_.Add(alias);
 | 
| -  }
 | 
| -
 | 
| -  const Place* CanonicalizeAlias(const Place& alias) {
 | 
| -    const Place* canonical = aliases_map_.Lookup(&alias);
 | 
| -    if (canonical == NULL) {
 | 
| -      canonical = Place::Wrap(isolate_,
 | 
| -                              alias,
 | 
| -                              kAnyInstanceAnyIndexAlias + aliases_.length());
 | 
| -      InsertAlias(canonical);
 | 
| -    }
 | 
| -    return canonical;
 | 
| -  }
 | 
| -
 | 
| -  BitVector* GetRepresentativesSet(intptr_t alias) {
 | 
| -    return (alias < representatives_.length()) ? representatives_[alias] : NULL;
 | 
| -  }
 | 
| -
 | 
| -  BitVector* EnsureSet(GrowableArray<BitVector*>* sets,
 | 
| -                       intptr_t alias) {
 | 
| -    while (sets->length() <= alias) {
 | 
| -      sets->Add(NULL);
 | 
| -    }
 | 
| -
 | 
| -    BitVector* set = (*sets)[alias];
 | 
| -    if (set == NULL) {
 | 
| -      (*sets)[alias] = set = new(isolate_) BitVector(isolate_, max_place_id());
 | 
| -    }
 | 
| -    return set;
 | 
| -  }
 | 
| -
 | 
| -  void AddAllRepresentatives(const Place* to, intptr_t from) {
 | 
| -    AddAllRepresentatives(to->id(), from);
 | 
| -  }
 | 
| -
 | 
| -  void AddAllRepresentatives(intptr_t to, intptr_t from) {
 | 
| -    BitVector* from_set = GetRepresentativesSet(from);
 | 
| -    if (from_set != NULL) {
 | 
| -      EnsureSet(&killed_, to)->AddAll(from_set);
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  void CrossAlias(const Place* to, const Place& from) {
 | 
| -    const intptr_t from_id = LookupAliasId(from);
 | 
| -    if (from_id == kNoAlias) {
 | 
| -      return;
 | 
| -    }
 | 
| -    CrossAlias(to, from_id);
 | 
| -  }
 | 
| -
 | 
| -  void CrossAlias(const Place* to, intptr_t from) {
 | 
| -    AddAllRepresentatives(to->id(), from);
 | 
| -    AddAllRepresentatives(from, to->id());
 | 
| -  }
 | 
| -
 | 
| -  // When computing kill sets we let less generic alias insert its
 | 
| -  // representatives into more generic alias'es kill set. For example
 | 
| -  // when visiting alias X[*] instead of searching for all aliases X[C]
 | 
| -  // and inserting their representatives into kill set for X[*] we update
 | 
| -  // kill set for X[*] each time we visit new X[C] for some C.
 | 
| -  // There is an exception however: if both aliases are parametric like *[C]
 | 
| -  // and X[*] which cross alias when X is an aliased allocation then we use
 | 
| -  // artificial aliases that contain all possible representatives for the given
 | 
| -  // alias for any value of the parameter to compute resulting kill set.
 | 
| -  void ComputeKillSet(const Place* alias) {
 | 
| -    switch (alias->kind()) {
 | 
| -      case Place::kIndexed:  // Either *[*] or X[*] alias.
 | 
| -        if (alias->instance() == NULL) {
 | 
| -          // *[*] aliases with X[*], X[C], *[C].
 | 
| -          AddAllRepresentatives(alias, kAnyConstantIndexedAlias);
 | 
| -          AddAllRepresentatives(alias, kAnyAllocationIndexedAlias);
 | 
| -        } else if (CanBeAliased(alias->instance())) {
 | 
| -          // X[*] aliases with X[C].
 | 
| -          // If X can be aliased then X[*] also aliases with *[C], *[*].
 | 
| -          CrossAlias(alias, kAnyInstanceAnyIndexAlias);
 | 
| -          AddAllRepresentatives(alias, kUnknownInstanceConstantIndexedAlias);
 | 
| -        }
 | 
| -        break;
 | 
| -
 | 
| -      case Place::kConstantIndexed:  // Either X[C] or *[C] alias.
 | 
| -        if (alias->instance() == NULL) {
 | 
| -          // *[C] aliases with X[C], X[*], *[*].
 | 
| -          AddAllRepresentatives(alias, kAnyAllocationIndexedAlias);
 | 
| -          CrossAlias(alias, kAnyInstanceAnyIndexAlias);
 | 
| -        } else {
 | 
| -          // X[C] aliases with X[*].
 | 
| -          // If X can be aliased then X[C] also aliases with *[C], *[*].
 | 
| -          CrossAlias(alias, alias->CopyWithoutIndex());
 | 
| -          if (CanBeAliased(alias->instance())) {
 | 
| -            CrossAlias(alias, alias->CopyWithoutInstance());
 | 
| -            CrossAlias(alias, kAnyInstanceAnyIndexAlias);
 | 
| -          }
 | 
| -        }
 | 
| -        break;
 | 
| -
 | 
| -      case Place::kField:
 | 
| -      case Place::kVMField:
 | 
| -        if (CanBeAliased(alias->instance())) {
 | 
| -          // X.f or X.@offs alias with *.f and *.@offs respectively.
 | 
| -          CrossAlias(alias, alias->CopyWithoutInstance());
 | 
| -        }
 | 
| -        break;
 | 
| -
 | 
| -      case Place::kNone:
 | 
| -        UNREACHABLE();
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Returns true if the given load is unaffected by external side-effects.
 | 
| -  // This essentially means that no stores to the same location can
 | 
| -  // occur in other functions.
 | 
| -  bool IsIndependentFromEffects(Place* place) {
 | 
| -    if (place->IsFinalField()) {
 | 
| -      // Note that we can't use LoadField's is_immutable attribute here because
 | 
| -      // some VM-fields (those that have no corresponding Field object and
 | 
| -      // accessed through offset alone) can share offset but have different
 | 
| -      // immutability properties.
 | 
| -      // One example is the length property of growable and fixed size list. If
 | 
| -      // loads of these two properties occur in the same function for the same
 | 
| -      // receiver then they will get the same expression number. However
 | 
| -      // immutability of the length of fixed size list does not mean that
 | 
| -      // growable list also has immutable property. Thus we will make a
 | 
| -      // conservative assumption for the VM-properties.
 | 
| -      // TODO(vegorov): disambiguate immutable and non-immutable VM-fields with
 | 
| -      // the same offset e.g. through recognized kind.
 | 
| -      return true;
 | 
| -    }
 | 
| -
 | 
| -    return ((place->kind() == Place::kField) ||
 | 
| -         (place->kind() == Place::kVMField)) &&
 | 
| -        !CanBeAliased(place->instance());
 | 
| -  }
 | 
| -
 | 
| -  // Returns true if there are direct loads from the given place.
 | 
| -  bool HasLoadsFromPlace(Definition* defn, const Place* place) {
 | 
| -    ASSERT((place->kind() == Place::kField) ||
 | 
| -           (place->kind() == Place::kVMField));
 | 
| -
 | 
| -    for (Value* use = defn->input_use_list();
 | 
| -         use != NULL;
 | 
| -         use = use->next_use()) {
 | 
| -      Instruction* instr = use->instruction();
 | 
| -      if ((instr->IsRedefinition() ||
 | 
| -           instr->IsAssertAssignable()) &&
 | 
| -          HasLoadsFromPlace(instr->AsDefinition(), place)) {
 | 
| -        return true;
 | 
| -      }
 | 
| -      bool is_load = false, is_store;
 | 
| -      Place load_place(instr, &is_load, &is_store);
 | 
| -
 | 
| -      if (is_load && load_place.Equals(place)) {
 | 
| -        return true;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  // Check if any use of the definition can create an alias.
 | 
| -  // Can add more objects into aliasing_worklist_.
 | 
| -  bool AnyUseCreatesAlias(Definition* defn) {
 | 
| -    for (Value* use = defn->input_use_list();
 | 
| -         use != NULL;
 | 
| -         use = use->next_use()) {
 | 
| -      Instruction* instr = use->instruction();
 | 
| -      if (instr->IsPushArgument() ||
 | 
| -          (instr->IsStoreIndexed()
 | 
| -           && (use->use_index() == StoreIndexedInstr::kValuePos)) ||
 | 
| -          instr->IsStoreStaticField() ||
 | 
| -          instr->IsPhi()) {
 | 
| -        return true;
 | 
| -      } else if ((instr->IsAssertAssignable() || instr->IsRedefinition()) &&
 | 
| -                 AnyUseCreatesAlias(instr->AsDefinition())) {
 | 
| -        return true;
 | 
| -      } else if ((instr->IsStoreInstanceField()
 | 
| -           && (use->use_index() != StoreInstanceFieldInstr::kInstancePos))) {
 | 
| -        ASSERT(use->use_index() == StoreInstanceFieldInstr::kValuePos);
 | 
| -        // If we store this value into an object that is not aliased itself
 | 
| -        // and we never load again then the store does not create an alias.
 | 
| -        StoreInstanceFieldInstr* store = instr->AsStoreInstanceField();
 | 
| -        Definition* instance =
 | 
| -            store->instance()->definition()->OriginalDefinition();
 | 
| -        if (Place::IsAllocation(instance) &&
 | 
| -            !instance->Identity().IsAliased()) {
 | 
| -          bool is_load, is_store;
 | 
| -          Place store_place(instr, &is_load, &is_store);
 | 
| -
 | 
| -          if (!HasLoadsFromPlace(instance, &store_place)) {
 | 
| -            // No loads found that match this store. If it is yet unknown if
 | 
| -            // the object is not aliased then optimistically assume this but
 | 
| -            // add it to the worklist to check its uses transitively.
 | 
| -            if (instance->Identity().IsUnknown()) {
 | 
| -              instance->SetIdentity(AliasIdentity::NotAliased());
 | 
| -              aliasing_worklist_.Add(instance);
 | 
| -            }
 | 
| -            continue;
 | 
| -          }
 | 
| -        }
 | 
| -        return true;
 | 
| -      }
 | 
| -    }
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  // Mark any value stored into the given object as potentially aliased.
 | 
| -  void MarkStoredValuesEscaping(Definition* defn) {
 | 
| -    // Find all stores into this object.
 | 
| -    for (Value* use = defn->input_use_list();
 | 
| -         use != NULL;
 | 
| -         use = use->next_use()) {
 | 
| -      if (use->instruction()->IsRedefinition() ||
 | 
| -          use->instruction()->IsAssertAssignable()) {
 | 
| -        MarkStoredValuesEscaping(use->instruction()->AsDefinition());
 | 
| -        continue;
 | 
| -      }
 | 
| -      if ((use->use_index() == StoreInstanceFieldInstr::kInstancePos) &&
 | 
| -          use->instruction()->IsStoreInstanceField()) {
 | 
| -        StoreInstanceFieldInstr* store =
 | 
| -            use->instruction()->AsStoreInstanceField();
 | 
| -        Definition* value = store->value()->definition()->OriginalDefinition();
 | 
| -        if (value->Identity().IsNotAliased()) {
 | 
| -          value->SetIdentity(AliasIdentity::Aliased());
 | 
| -          identity_rollback_.Add(value);
 | 
| -
 | 
| -          // Add to worklist to propagate the mark transitively.
 | 
| -          aliasing_worklist_.Add(value);
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Determine if the given definition can't be aliased.
 | 
| -  void ComputeAliasing(Definition* alloc) {
 | 
| -    ASSERT(Place::IsAllocation(alloc));
 | 
| -    ASSERT(alloc->Identity().IsUnknown());
 | 
| -    ASSERT(aliasing_worklist_.is_empty());
 | 
| -
 | 
| -    alloc->SetIdentity(AliasIdentity::NotAliased());
 | 
| -    aliasing_worklist_.Add(alloc);
 | 
| -
 | 
| -    while (!aliasing_worklist_.is_empty()) {
 | 
| -      Definition* defn = aliasing_worklist_.RemoveLast();
 | 
| -      ASSERT(Place::IsAllocation(defn));
 | 
| -      // If the definition in the worklist was optimistically marked as
 | 
| -      // not-aliased check that optimistic assumption still holds: check if
 | 
| -      // any of its uses can create an alias.
 | 
| -      if (!defn->Identity().IsAliased() && AnyUseCreatesAlias(defn)) {
 | 
| -        defn->SetIdentity(AliasIdentity::Aliased());
 | 
| -        identity_rollback_.Add(defn);
 | 
| -      }
 | 
| -
 | 
| -      // If the allocation site is marked as aliased conservatively mark
 | 
| -      // any values stored into the object aliased too.
 | 
| -      if (defn->Identity().IsAliased()) {
 | 
| -        MarkStoredValuesEscaping(defn);
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  Isolate* isolate_;
 | 
| -
 | 
| -  DirectChainedHashMap<PointerKeyValueTrait<Place> >* places_map_;
 | 
| -
 | 
| -  const ZoneGrowableArray<Place*>& places_;
 | 
| -
 | 
| -  const PhiPlaceMoves* phi_moves_;
 | 
| -
 | 
| -  // A list of all seen aliases and a map that allows looking up canonical
 | 
| -  // alias object.
 | 
| -  GrowableArray<const Place*> aliases_;
 | 
| -  DirectChainedHashMap<PointerKeyValueTrait<const Place> > aliases_map_;
 | 
| -
 | 
| -  // Maps alias id to set of ids of places representing the alias.
 | 
| -  // Place represents an alias if this alias is least generic alias for
 | 
| -  // the place.
 | 
| -  // (see ToAlias for the definition of least generic alias).
 | 
| -  GrowableArray<BitVector*> representatives_;
 | 
| -
 | 
| -  // Maps alias id to set of ids of places aliased.
 | 
| -  GrowableArray<BitVector*> killed_;
 | 
| -
 | 
| -  // Set of ids of places that can be affected by side-effects other than
 | 
| -  // explicit stores (i.e. through calls).
 | 
| -  BitVector* aliased_by_effects_;
 | 
| -
 | 
| -  // Worklist used during alias analysis.
 | 
| -  GrowableArray<Definition*> aliasing_worklist_;
 | 
| -
 | 
| -  // List of definitions that had their identity set to Aliased. At the end
 | 
| -  // of load optimization their identity will be rolled back to Unknown to
 | 
| -  // avoid treating them as Aliased at later stages without checking first
 | 
| -  // as optimizations can potentially eliminate instructions leading to
 | 
| -  // aliasing.
 | 
| -  GrowableArray<Definition*> identity_rollback_;
 | 
| -};
 | 
| -
 | 
| -
 | 
| -static Definition* GetStoredValue(Instruction* instr) {
 | 
| -  if (instr->IsStoreIndexed()) {
 | 
| -    return instr->AsStoreIndexed()->value()->definition();
 | 
| -  }
 | 
| -
 | 
| -  StoreInstanceFieldInstr* store_instance_field = instr->AsStoreInstanceField();
 | 
| -  if (store_instance_field != NULL) {
 | 
| -    return store_instance_field->value()->definition();
 | 
| -  }
 | 
| -
 | 
| -  StoreStaticFieldInstr* store_static_field = instr->AsStoreStaticField();
 | 
| -  if (store_static_field != NULL) {
 | 
| -    return store_static_field->value()->definition();
 | 
| -  }
 | 
| -
 | 
| -  UNREACHABLE();  // Should only be called for supported store instructions.
 | 
| -  return NULL;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool IsPhiDependentPlace(Place* place) {
 | 
| -  return ((place->kind() == Place::kField) ||
 | 
| -          (place->kind() == Place::kVMField)) &&
 | 
| -        (place->instance() != NULL) &&
 | 
| -        place->instance()->IsPhi();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// For each place that depends on a phi ensure that equivalent places
 | 
| -// corresponding to phi input are numbered and record outgoing phi moves
 | 
| -// for each block which establish correspondence between phi dependent place
 | 
| -// and phi input's place that is flowing in.
 | 
| -static PhiPlaceMoves* ComputePhiMoves(
 | 
| -    DirectChainedHashMap<PointerKeyValueTrait<Place> >* map,
 | 
| -    ZoneGrowableArray<Place*>* places) {
 | 
| -  Isolate* isolate = Isolate::Current();
 | 
| -  PhiPlaceMoves* phi_moves = new(isolate) PhiPlaceMoves();
 | 
| -
 | 
| -  for (intptr_t i = 0; i < places->length(); i++) {
 | 
| -    Place* place = (*places)[i];
 | 
| -
 | 
| -    if (IsPhiDependentPlace(place)) {
 | 
| -      PhiInstr* phi = place->instance()->AsPhi();
 | 
| -      BlockEntryInstr* block = phi->GetBlock();
 | 
| -
 | 
| -      if (FLAG_trace_optimization) {
 | 
| -        OS::Print("phi dependent place %s\n", place->ToCString());
 | 
| -      }
 | 
| -
 | 
| -      Place input_place(*place);
 | 
| -      for (intptr_t j = 0; j < phi->InputCount(); j++) {
 | 
| -        input_place.set_instance(phi->InputAt(j)->definition());
 | 
| -
 | 
| -        Place* result = map->Lookup(&input_place);
 | 
| -        if (result == NULL) {
 | 
| -          result = Place::Wrap(isolate, input_place, places->length());
 | 
| -          map->Insert(result);
 | 
| -          places->Add(result);
 | 
| -          if (FLAG_trace_optimization) {
 | 
| -            OS::Print("  adding place %s as %" Pd "\n",
 | 
| -                      result->ToCString(),
 | 
| -                      result->id());
 | 
| -          }
 | 
| -        }
 | 
| -        phi_moves->CreateOutgoingMove(isolate,
 | 
| -                                      block->PredecessorAt(j),
 | 
| -                                      result->id(),
 | 
| -                                      place->id());
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  return phi_moves;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -enum CSEMode {
 | 
| -  kOptimizeLoads,
 | 
| -  kOptimizeStores
 | 
| -};
 | 
| -
 | 
| -
 | 
| -static AliasedSet* NumberPlaces(
 | 
| -    FlowGraph* graph,
 | 
| -    DirectChainedHashMap<PointerKeyValueTrait<Place> >* map,
 | 
| -    CSEMode mode) {
 | 
| -  // Loads representing different expression ids will be collected and
 | 
| -  // used to build per offset kill sets.
 | 
| -  Isolate* isolate = graph->isolate();
 | 
| -  ZoneGrowableArray<Place*>* places =
 | 
| -      new(isolate) ZoneGrowableArray<Place*>(10);
 | 
| -
 | 
| -  bool has_loads = false;
 | 
| -  bool has_stores = false;
 | 
| -  for (BlockIterator it = graph->reverse_postorder_iterator();
 | 
| -       !it.Done();
 | 
| -       it.Advance()) {
 | 
| -    BlockEntryInstr* block = it.Current();
 | 
| -
 | 
| -    for (ForwardInstructionIterator instr_it(block);
 | 
| -         !instr_it.Done();
 | 
| -         instr_it.Advance()) {
 | 
| -      Instruction* instr = instr_it.Current();
 | 
| -      Place place(instr, &has_loads, &has_stores);
 | 
| -      if (place.kind() == Place::kNone) {
 | 
| -        continue;
 | 
| -      }
 | 
| -
 | 
| -      Place* result = map->Lookup(&place);
 | 
| -      if (result == NULL) {
 | 
| -        result = Place::Wrap(isolate, place, places->length());
 | 
| -        map->Insert(result);
 | 
| -        places->Add(result);
 | 
| -
 | 
| -        if (FLAG_trace_optimization) {
 | 
| -          OS::Print("numbering %s as %" Pd "\n",
 | 
| -                    result->ToCString(),
 | 
| -                    result->id());
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      instr->set_place_id(result->id());
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if ((mode == kOptimizeLoads) && !has_loads) {
 | 
| -    return NULL;
 | 
| -  }
 | 
| -  if ((mode == kOptimizeStores) && !has_stores) {
 | 
| -    return NULL;
 | 
| -  }
 | 
| -
 | 
| -  PhiPlaceMoves* phi_moves = ComputePhiMoves(map, places);
 | 
| -
 | 
| -  // Build aliasing sets mapping aliases to loads.
 | 
| -  return new(isolate) AliasedSet(isolate, map, places, phi_moves);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -class LoadOptimizer : public ValueObject {
 | 
| - public:
 | 
| -  LoadOptimizer(FlowGraph* graph, AliasedSet* aliased_set)
 | 
| -      : graph_(graph),
 | 
| -        aliased_set_(aliased_set),
 | 
| -        in_(graph_->preorder().length()),
 | 
| -        out_(graph_->preorder().length()),
 | 
| -        gen_(graph_->preorder().length()),
 | 
| -        kill_(graph_->preorder().length()),
 | 
| -        exposed_values_(graph_->preorder().length()),
 | 
| -        out_values_(graph_->preorder().length()),
 | 
| -        phis_(5),
 | 
| -        worklist_(5),
 | 
| -        congruency_worklist_(6),
 | 
| -        in_worklist_(NULL),
 | 
| -        forwarded_(false) {
 | 
| -    const intptr_t num_blocks = graph_->preorder().length();
 | 
| -    for (intptr_t i = 0; i < num_blocks; i++) {
 | 
| -      out_.Add(NULL);
 | 
| -      gen_.Add(new(I) BitVector(I, aliased_set_->max_place_id()));
 | 
| -      kill_.Add(new(I) BitVector(I, aliased_set_->max_place_id()));
 | 
| -      in_.Add(new(I) BitVector(I, aliased_set_->max_place_id()));
 | 
| -
 | 
| -      exposed_values_.Add(NULL);
 | 
| -      out_values_.Add(NULL);
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  ~LoadOptimizer() {
 | 
| -    aliased_set_->RollbackAliasedIdentites();
 | 
| -  }
 | 
| -
 | 
| -  Isolate* isolate() const { return graph_->isolate(); }
 | 
| -
 | 
| -  static bool OptimizeGraph(FlowGraph* graph) {
 | 
| -    ASSERT(FLAG_load_cse);
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      FlowGraphPrinter::PrintGraph("Before LoadOptimizer", graph);
 | 
| -    }
 | 
| -
 | 
| -    DirectChainedHashMap<PointerKeyValueTrait<Place> > map;
 | 
| -    AliasedSet* aliased_set = NumberPlaces(graph, &map, kOptimizeLoads);
 | 
| -    if ((aliased_set != NULL) && !aliased_set->IsEmpty()) {
 | 
| -      // If any loads were forwarded return true from Optimize to run load
 | 
| -      // forwarding again. This will allow to forward chains of loads.
 | 
| -      // This is especially important for context variables as they are built
 | 
| -      // as loads from loaded context.
 | 
| -      // TODO(vegorov): renumber newly discovered congruences during the
 | 
| -      // forwarding to forward chains without running whole pass twice.
 | 
| -      LoadOptimizer load_optimizer(graph, aliased_set);
 | 
| -      return load_optimizer.Optimize();
 | 
| -    }
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| - private:
 | 
| -  bool Optimize() {
 | 
| -    ComputeInitialSets();
 | 
| -    ComputeOutSets();
 | 
| -    ComputeOutValues();
 | 
| -    if (graph_->is_licm_allowed()) {
 | 
| -      MarkLoopInvariantLoads();
 | 
| -    }
 | 
| -    ForwardLoads();
 | 
| -    EmitPhis();
 | 
| -
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      FlowGraphPrinter::PrintGraph("After LoadOptimizer", graph_);
 | 
| -    }
 | 
| -
 | 
| -    return forwarded_;
 | 
| -  }
 | 
| -
 | 
| -  // Compute sets of loads generated and killed by each block.
 | 
| -  // Additionally compute upwards exposed and generated loads for each block.
 | 
| -  // Exposed loads are those that can be replaced if a corresponding
 | 
| -  // reaching load will be found.
 | 
| -  // Loads that are locally redundant will be replaced as we go through
 | 
| -  // instructions.
 | 
| -  void ComputeInitialSets() {
 | 
| -    for (BlockIterator block_it = graph_->reverse_postorder_iterator();
 | 
| -         !block_it.Done();
 | 
| -         block_it.Advance()) {
 | 
| -      BlockEntryInstr* block = block_it.Current();
 | 
| -      const intptr_t preorder_number = block->preorder_number();
 | 
| -
 | 
| -      BitVector* kill = kill_[preorder_number];
 | 
| -      BitVector* gen = gen_[preorder_number];
 | 
| -
 | 
| -      ZoneGrowableArray<Definition*>* exposed_values = NULL;
 | 
| -      ZoneGrowableArray<Definition*>* out_values = NULL;
 | 
| -
 | 
| -      for (ForwardInstructionIterator instr_it(block);
 | 
| -           !instr_it.Done();
 | 
| -           instr_it.Advance()) {
 | 
| -        Instruction* instr = instr_it.Current();
 | 
| -
 | 
| -        bool is_load = false, is_store = false;
 | 
| -        Place place(instr, &is_load, &is_store);
 | 
| -
 | 
| -        BitVector* killed = NULL;
 | 
| -        if (is_store) {
 | 
| -          const intptr_t alias_id =
 | 
| -              aliased_set_->LookupAliasId(place.ToAlias());
 | 
| -          if (alias_id != AliasedSet::kNoAlias) {
 | 
| -            killed = aliased_set_->GetKilledSet(alias_id);
 | 
| -          } else if (!place.IsFinalField()) {
 | 
| -            // We encountered unknown alias: this means intrablock load
 | 
| -            // forwarding refined parameter of this store, for example
 | 
| -            //
 | 
| -            //     o   <- alloc()
 | 
| -            //     a.f <- o
 | 
| -            //     u   <- a.f
 | 
| -            //     u.x <- null ;; this store alias is *.x
 | 
| -            //
 | 
| -            // after intrablock load forwarding
 | 
| -            //
 | 
| -            //     o   <- alloc()
 | 
| -            //     a.f <- o
 | 
| -            //     o.x <- null ;; this store alias is o.x
 | 
| -            //
 | 
| -            // In this case we fallback to using place id recorded in the
 | 
| -            // instruction that still points to the old place with a more
 | 
| -            // generic alias.
 | 
| -            const intptr_t old_alias_id = aliased_set_->LookupAliasId(
 | 
| -                aliased_set_->places()[instr->place_id()]->ToAlias());
 | 
| -            killed = aliased_set_->GetKilledSet(old_alias_id);
 | 
| -          }
 | 
| -
 | 
| -          if (killed != NULL) {
 | 
| -            kill->AddAll(killed);
 | 
| -            // There is no need to clear out_values when clearing GEN set
 | 
| -            // because only those values that are in the GEN set
 | 
| -            // will ever be used.
 | 
| -            gen->RemoveAll(killed);
 | 
| -          }
 | 
| -
 | 
| -          // Only forward stores to normal arrays, float64, and simd arrays
 | 
| -          // to loads because other array stores (intXX/uintXX/float32)
 | 
| -          // may implicitly convert the value stored.
 | 
| -          StoreIndexedInstr* array_store = instr->AsStoreIndexed();
 | 
| -          if ((array_store == NULL) ||
 | 
| -              (array_store->class_id() == kArrayCid) ||
 | 
| -              (array_store->class_id() == kTypedDataFloat64ArrayCid) ||
 | 
| -              (array_store->class_id() == kTypedDataFloat32ArrayCid) ||
 | 
| -              (array_store->class_id() == kTypedDataFloat32x4ArrayCid)) {
 | 
| -            Place* canonical_place = aliased_set_->LookupCanonical(&place);
 | 
| -            if (canonical_place != NULL) {
 | 
| -              // Store has a corresponding numbered place that might have a
 | 
| -              // load. Try forwarding stored value to it.
 | 
| -              gen->Add(canonical_place->id());
 | 
| -              if (out_values == NULL) out_values = CreateBlockOutValues();
 | 
| -              (*out_values)[canonical_place->id()] = GetStoredValue(instr);
 | 
| -            }
 | 
| -          }
 | 
| -
 | 
| -          ASSERT(!instr->IsDefinition() ||
 | 
| -                 !IsLoadEliminationCandidate(instr->AsDefinition()));
 | 
| -          continue;
 | 
| -        } else if (is_load) {
 | 
| -          // Check if this load needs renumbering because of the intrablock
 | 
| -          // load forwarding.
 | 
| -          const Place* canonical = aliased_set_->LookupCanonical(&place);
 | 
| -          if ((canonical != NULL) &&
 | 
| -            (canonical->id() != instr->AsDefinition()->place_id())) {
 | 
| -            instr->AsDefinition()->set_place_id(canonical->id());
 | 
| -          }
 | 
| -        }
 | 
| -
 | 
| -        // If instruction has effects then kill all loads affected.
 | 
| -        if (!instr->Effects().IsNone()) {
 | 
| -          kill->AddAll(aliased_set_->aliased_by_effects());
 | 
| -          // There is no need to clear out_values when removing values from GEN
 | 
| -          // set because only those values that are in the GEN set
 | 
| -          // will ever be used.
 | 
| -          gen->RemoveAll(aliased_set_->aliased_by_effects());
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        Definition* defn = instr->AsDefinition();
 | 
| -        if (defn == NULL) {
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        // For object allocation forward initial values of the fields to
 | 
| -        // subsequent loads. For skip final fields.  Final fields are
 | 
| -        // initialized in constructor that potentially can be not inlined into
 | 
| -        // the function that we are currently optimizing. However at the same
 | 
| -        // time we assume that values of the final fields can be forwarded
 | 
| -        // across side-effects. If we add 'null' as known values for these
 | 
| -        // fields here we will incorrectly propagate this null across
 | 
| -        // constructor invocation.
 | 
| -        AllocateObjectInstr* alloc = instr->AsAllocateObject();
 | 
| -        if ((alloc != NULL)) {
 | 
| -          for (Value* use = alloc->input_use_list();
 | 
| -               use != NULL;
 | 
| -               use = use->next_use()) {
 | 
| -            // Look for all immediate loads from this object.
 | 
| -            if (use->use_index() != 0) {
 | 
| -              continue;
 | 
| -            }
 | 
| -
 | 
| -            LoadFieldInstr* load = use->instruction()->AsLoadField();
 | 
| -            if (load != NULL) {
 | 
| -              // Found a load. Initialize current value of the field to null for
 | 
| -              // normal fields, or with type arguments.
 | 
| -
 | 
| -              // Forward for all fields for non-escaping objects and only
 | 
| -              // non-final fields and type arguments for escaping ones.
 | 
| -              if (aliased_set_->CanBeAliased(alloc) &&
 | 
| -                  (load->field() != NULL) &&
 | 
| -                  load->field()->is_final()) {
 | 
| -                continue;
 | 
| -              }
 | 
| -
 | 
| -              Definition* forward_def = graph_->constant_null();
 | 
| -              if (alloc->ArgumentCount() > 0) {
 | 
| -                ASSERT(alloc->ArgumentCount() == 1);
 | 
| -                intptr_t type_args_offset =
 | 
| -                    alloc->cls().type_arguments_field_offset();
 | 
| -                if (load->offset_in_bytes() == type_args_offset) {
 | 
| -                  forward_def = alloc->PushArgumentAt(0)->value()->definition();
 | 
| -                }
 | 
| -              }
 | 
| -              gen->Add(load->place_id());
 | 
| -              if (out_values == NULL) out_values = CreateBlockOutValues();
 | 
| -              (*out_values)[load->place_id()] = forward_def;
 | 
| -            }
 | 
| -          }
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        if (!IsLoadEliminationCandidate(defn)) {
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        const intptr_t place_id = defn->place_id();
 | 
| -        if (gen->Contains(place_id)) {
 | 
| -          // This is a locally redundant load.
 | 
| -          ASSERT((out_values != NULL) && ((*out_values)[place_id] != NULL));
 | 
| -
 | 
| -          Definition* replacement = (*out_values)[place_id];
 | 
| -          EnsureSSATempIndex(graph_, defn, replacement);
 | 
| -          if (FLAG_trace_optimization) {
 | 
| -            OS::Print("Replacing load v%" Pd " with v%" Pd "\n",
 | 
| -                      defn->ssa_temp_index(),
 | 
| -                      replacement->ssa_temp_index());
 | 
| -          }
 | 
| -
 | 
| -          defn->ReplaceUsesWith(replacement);
 | 
| -          instr_it.RemoveCurrentFromGraph();
 | 
| -          forwarded_ = true;
 | 
| -          continue;
 | 
| -        } else if (!kill->Contains(place_id)) {
 | 
| -          // This is an exposed load: it is the first representative of a
 | 
| -          // given expression id and it is not killed on the path from
 | 
| -          // the block entry.
 | 
| -          if (exposed_values == NULL) {
 | 
| -            static const intptr_t kMaxExposedValuesInitialSize = 5;
 | 
| -            exposed_values = new(I) ZoneGrowableArray<Definition*>(
 | 
| -                Utils::Minimum(kMaxExposedValuesInitialSize,
 | 
| -                               aliased_set_->max_place_id()));
 | 
| -          }
 | 
| -
 | 
| -          exposed_values->Add(defn);
 | 
| -        }
 | 
| -
 | 
| -        gen->Add(place_id);
 | 
| -
 | 
| -        if (out_values == NULL) out_values = CreateBlockOutValues();
 | 
| -        (*out_values)[place_id] = defn;
 | 
| -      }
 | 
| -
 | 
| -      exposed_values_[preorder_number] = exposed_values;
 | 
| -      out_values_[preorder_number] = out_values;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  static void PerformPhiMoves(PhiPlaceMoves::MovesList phi_moves,
 | 
| -                              BitVector* out,
 | 
| -                              BitVector* forwarded_loads) {
 | 
| -    forwarded_loads->Clear();
 | 
| -
 | 
| -    for (intptr_t i = 0; i < phi_moves->length(); i++) {
 | 
| -      const intptr_t from = (*phi_moves)[i].from();
 | 
| -      const intptr_t to = (*phi_moves)[i].to();
 | 
| -      if (from == to) continue;
 | 
| -
 | 
| -      if (out->Contains(from)) {
 | 
| -        forwarded_loads->Add(to);
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    for (intptr_t i = 0; i < phi_moves->length(); i++) {
 | 
| -      const intptr_t from = (*phi_moves)[i].from();
 | 
| -      const intptr_t to = (*phi_moves)[i].to();
 | 
| -      if (from == to) continue;
 | 
| -
 | 
| -      out->Remove(to);
 | 
| -    }
 | 
| -
 | 
| -    out->AddAll(forwarded_loads);
 | 
| -  }
 | 
| -
 | 
| -  // Compute OUT sets by propagating them iteratively until fix point
 | 
| -  // is reached.
 | 
| -  void ComputeOutSets() {
 | 
| -    BitVector* temp = new(I) BitVector(I, aliased_set_->max_place_id());
 | 
| -    BitVector* forwarded_loads =
 | 
| -        new(I) BitVector(I, aliased_set_->max_place_id());
 | 
| -    BitVector* temp_out = new(I) BitVector(I, aliased_set_->max_place_id());
 | 
| -
 | 
| -    bool changed = true;
 | 
| -    while (changed) {
 | 
| -      changed = false;
 | 
| -
 | 
| -      for (BlockIterator block_it = graph_->reverse_postorder_iterator();
 | 
| -           !block_it.Done();
 | 
| -           block_it.Advance()) {
 | 
| -        BlockEntryInstr* block = block_it.Current();
 | 
| -
 | 
| -        const intptr_t preorder_number = block->preorder_number();
 | 
| -
 | 
| -        BitVector* block_in = in_[preorder_number];
 | 
| -        BitVector* block_out = out_[preorder_number];
 | 
| -        BitVector* block_kill = kill_[preorder_number];
 | 
| -        BitVector* block_gen = gen_[preorder_number];
 | 
| -
 | 
| -        // Compute block_in as the intersection of all out(p) where p
 | 
| -        // is a predecessor of the current block.
 | 
| -        if (block->IsGraphEntry()) {
 | 
| -          temp->Clear();
 | 
| -        } else {
 | 
| -          temp->SetAll();
 | 
| -          ASSERT(block->PredecessorCount() > 0);
 | 
| -          for (intptr_t i = 0; i < block->PredecessorCount(); i++) {
 | 
| -            BlockEntryInstr* pred = block->PredecessorAt(i);
 | 
| -            BitVector* pred_out = out_[pred->preorder_number()];
 | 
| -            if (pred_out == NULL) continue;
 | 
| -            PhiPlaceMoves::MovesList phi_moves =
 | 
| -                aliased_set_->phi_moves()->GetOutgoingMoves(pred);
 | 
| -            if (phi_moves != NULL) {
 | 
| -              // If there are phi moves, perform intersection with
 | 
| -              // a copy of pred_out where the phi moves are applied.
 | 
| -              temp_out->CopyFrom(pred_out);
 | 
| -              PerformPhiMoves(phi_moves, temp_out, forwarded_loads);
 | 
| -              pred_out = temp_out;
 | 
| -            }
 | 
| -            temp->Intersect(pred_out);
 | 
| -          }
 | 
| -        }
 | 
| -
 | 
| -        if (!temp->Equals(*block_in) || (block_out == NULL)) {
 | 
| -          // If IN set has changed propagate the change to OUT set.
 | 
| -          block_in->CopyFrom(temp);
 | 
| -
 | 
| -          temp->RemoveAll(block_kill);
 | 
| -          temp->AddAll(block_gen);
 | 
| -
 | 
| -          if ((block_out == NULL) || !block_out->Equals(*temp)) {
 | 
| -            if (block_out == NULL) {
 | 
| -              block_out = out_[preorder_number] =
 | 
| -                  new(I) BitVector(I, aliased_set_->max_place_id());
 | 
| -            }
 | 
| -            block_out->CopyFrom(temp);
 | 
| -            changed = true;
 | 
| -          }
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Compute out_values mappings by propagating them in reverse postorder once
 | 
| -  // through the graph. Generate phis on back edges where eager merge is
 | 
| -  // impossible.
 | 
| -  // No replacement is done at this point and thus any out_value[place_id] is
 | 
| -  // changed at most once: from NULL to an actual value.
 | 
| -  // When merging incoming loads we might need to create a phi.
 | 
| -  // These phis are not inserted at the graph immediately because some of them
 | 
| -  // might become redundant after load forwarding is done.
 | 
| -  void ComputeOutValues() {
 | 
| -    GrowableArray<PhiInstr*> pending_phis(5);
 | 
| -    ZoneGrowableArray<Definition*>* temp_forwarded_values = NULL;
 | 
| -
 | 
| -    for (BlockIterator block_it = graph_->reverse_postorder_iterator();
 | 
| -         !block_it.Done();
 | 
| -         block_it.Advance()) {
 | 
| -      BlockEntryInstr* block = block_it.Current();
 | 
| -
 | 
| -      const bool can_merge_eagerly = CanMergeEagerly(block);
 | 
| -
 | 
| -      const intptr_t preorder_number = block->preorder_number();
 | 
| -
 | 
| -      ZoneGrowableArray<Definition*>* block_out_values =
 | 
| -          out_values_[preorder_number];
 | 
| -
 | 
| -
 | 
| -      // If OUT set has changed then we have new values available out of
 | 
| -      // the block. Compute these values creating phi where necessary.
 | 
| -      for (BitVector::Iterator it(out_[preorder_number]);
 | 
| -           !it.Done();
 | 
| -           it.Advance()) {
 | 
| -        const intptr_t place_id = it.Current();
 | 
| -
 | 
| -        if (block_out_values == NULL) {
 | 
| -          out_values_[preorder_number] = block_out_values =
 | 
| -              CreateBlockOutValues();
 | 
| -        }
 | 
| -
 | 
| -        if ((*block_out_values)[place_id] == NULL) {
 | 
| -          ASSERT(block->PredecessorCount() > 0);
 | 
| -          Definition* in_value = can_merge_eagerly ?
 | 
| -              MergeIncomingValues(block, place_id) : NULL;
 | 
| -          if ((in_value == NULL) &&
 | 
| -              (in_[preorder_number]->Contains(place_id))) {
 | 
| -            PhiInstr* phi = new(I) PhiInstr(block->AsJoinEntry(),
 | 
| -                                            block->PredecessorCount());
 | 
| -            phi->set_place_id(place_id);
 | 
| -            pending_phis.Add(phi);
 | 
| -            in_value = phi;
 | 
| -          }
 | 
| -          (*block_out_values)[place_id] = in_value;
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      // If the block has outgoing phi moves perform them. Use temporary list
 | 
| -      // of values to ensure that cyclic moves are performed correctly.
 | 
| -      PhiPlaceMoves::MovesList phi_moves =
 | 
| -          aliased_set_->phi_moves()->GetOutgoingMoves(block);
 | 
| -      if ((phi_moves != NULL) && (block_out_values != NULL)) {
 | 
| -        if (temp_forwarded_values == NULL) {
 | 
| -          temp_forwarded_values = CreateBlockOutValues();
 | 
| -        }
 | 
| -
 | 
| -        for (intptr_t i = 0; i < phi_moves->length(); i++) {
 | 
| -          const intptr_t from = (*phi_moves)[i].from();
 | 
| -          const intptr_t to = (*phi_moves)[i].to();
 | 
| -          if (from == to) continue;
 | 
| -
 | 
| -          (*temp_forwarded_values)[to] = (*block_out_values)[from];
 | 
| -        }
 | 
| -
 | 
| -        for (intptr_t i = 0; i < phi_moves->length(); i++) {
 | 
| -          const intptr_t from = (*phi_moves)[i].from();
 | 
| -          const intptr_t to = (*phi_moves)[i].to();
 | 
| -          if (from == to) continue;
 | 
| -
 | 
| -          (*block_out_values)[to] = (*temp_forwarded_values)[to];
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      if (FLAG_trace_load_optimization) {
 | 
| -        OS::Print("B%" Pd "\n", block->block_id());
 | 
| -        OS::Print("  IN: ");
 | 
| -        aliased_set_->PrintSet(in_[preorder_number]);
 | 
| -        OS::Print("\n");
 | 
| -
 | 
| -        OS::Print("  KILL: ");
 | 
| -        aliased_set_->PrintSet(kill_[preorder_number]);
 | 
| -        OS::Print("\n");
 | 
| -
 | 
| -        OS::Print("  OUT: ");
 | 
| -        aliased_set_->PrintSet(out_[preorder_number]);
 | 
| -        OS::Print("\n");
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    // All blocks were visited. Fill pending phis with inputs
 | 
| -    // that flow on back edges.
 | 
| -    for (intptr_t i = 0; i < pending_phis.length(); i++) {
 | 
| -      FillPhiInputs(pending_phis[i]);
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  bool CanMergeEagerly(BlockEntryInstr* block) {
 | 
| -    for (intptr_t i = 0; i < block->PredecessorCount(); i++) {
 | 
| -      BlockEntryInstr* pred = block->PredecessorAt(i);
 | 
| -      if (pred->postorder_number() < block->postorder_number()) {
 | 
| -        return false;
 | 
| -      }
 | 
| -    }
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  void MarkLoopInvariantLoads() {
 | 
| -    const ZoneGrowableArray<BlockEntryInstr*>& loop_headers =
 | 
| -        graph_->LoopHeaders();
 | 
| -
 | 
| -    ZoneGrowableArray<BitVector*>* invariant_loads =
 | 
| -        new(I) ZoneGrowableArray<BitVector*>(loop_headers.length());
 | 
| -
 | 
| -    for (intptr_t i = 0; i < loop_headers.length(); i++) {
 | 
| -      BlockEntryInstr* header = loop_headers[i];
 | 
| -      BlockEntryInstr* pre_header = header->ImmediateDominator();
 | 
| -      if (pre_header == NULL) {
 | 
| -        invariant_loads->Add(NULL);
 | 
| -        continue;
 | 
| -      }
 | 
| -
 | 
| -      BitVector* loop_gen = new(I) BitVector(I, aliased_set_->max_place_id());
 | 
| -      for (BitVector::Iterator loop_it(header->loop_info());
 | 
| -           !loop_it.Done();
 | 
| -           loop_it.Advance()) {
 | 
| -        const intptr_t preorder_number = loop_it.Current();
 | 
| -        loop_gen->AddAll(gen_[preorder_number]);
 | 
| -      }
 | 
| -
 | 
| -      for (BitVector::Iterator loop_it(header->loop_info());
 | 
| -           !loop_it.Done();
 | 
| -           loop_it.Advance()) {
 | 
| -        const intptr_t preorder_number = loop_it.Current();
 | 
| -        loop_gen->RemoveAll(kill_[preorder_number]);
 | 
| -      }
 | 
| -
 | 
| -      if (FLAG_trace_optimization) {
 | 
| -        for (BitVector::Iterator it(loop_gen); !it.Done(); it.Advance()) {
 | 
| -          OS::Print("place %s is loop invariant for B%" Pd "\n",
 | 
| -                    aliased_set_->places()[it.Current()]->ToCString(),
 | 
| -                    header->block_id());
 | 
| -        }
 | 
| -      }
 | 
| -
 | 
| -      invariant_loads->Add(loop_gen);
 | 
| -    }
 | 
| -
 | 
| -    graph_->set_loop_invariant_loads(invariant_loads);
 | 
| -  }
 | 
| -
 | 
| -  // Compute incoming value for the given expression id.
 | 
| -  // Will create a phi if different values are incoming from multiple
 | 
| -  // predecessors.
 | 
| -  Definition* MergeIncomingValues(BlockEntryInstr* block, intptr_t place_id) {
 | 
| -    // First check if the same value is coming in from all predecessors.
 | 
| -    static Definition* const kDifferentValuesMarker =
 | 
| -        reinterpret_cast<Definition*>(-1);
 | 
| -    Definition* incoming = NULL;
 | 
| -    for (intptr_t i = 0; i < block->PredecessorCount(); i++) {
 | 
| -      BlockEntryInstr* pred = block->PredecessorAt(i);
 | 
| -      ZoneGrowableArray<Definition*>* pred_out_values =
 | 
| -          out_values_[pred->preorder_number()];
 | 
| -      if ((pred_out_values == NULL) || ((*pred_out_values)[place_id] == NULL)) {
 | 
| -        return NULL;
 | 
| -      } else if (incoming == NULL) {
 | 
| -        incoming = (*pred_out_values)[place_id];
 | 
| -      } else if (incoming != (*pred_out_values)[place_id]) {
 | 
| -        incoming = kDifferentValuesMarker;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    if (incoming != kDifferentValuesMarker) {
 | 
| -      ASSERT(incoming != NULL);
 | 
| -      return incoming;
 | 
| -    }
 | 
| -
 | 
| -    // Incoming values are different. Phi is required to merge.
 | 
| -    PhiInstr* phi = new(I) PhiInstr(
 | 
| -        block->AsJoinEntry(), block->PredecessorCount());
 | 
| -    phi->set_place_id(place_id);
 | 
| -    FillPhiInputs(phi);
 | 
| -    return phi;
 | 
| -  }
 | 
| -
 | 
| -  void FillPhiInputs(PhiInstr* phi) {
 | 
| -    BlockEntryInstr* block = phi->GetBlock();
 | 
| -    const intptr_t place_id = phi->place_id();
 | 
| -
 | 
| -    for (intptr_t i = 0; i < block->PredecessorCount(); i++) {
 | 
| -      BlockEntryInstr* pred = block->PredecessorAt(i);
 | 
| -      ZoneGrowableArray<Definition*>* pred_out_values =
 | 
| -          out_values_[pred->preorder_number()];
 | 
| -      ASSERT((*pred_out_values)[place_id] != NULL);
 | 
| -
 | 
| -      // Sets of outgoing values are not linked into use lists so
 | 
| -      // they might contain values that were replaced and removed
 | 
| -      // from the graph by this iteration.
 | 
| -      // To prevent using them we additionally mark definitions themselves
 | 
| -      // as replaced and store a pointer to the replacement.
 | 
| -      Definition* replacement = (*pred_out_values)[place_id]->Replacement();
 | 
| -      Value* input = new(I) Value(replacement);
 | 
| -      phi->SetInputAt(i, input);
 | 
| -      replacement->AddInputUse(input);
 | 
| -    }
 | 
| -
 | 
| -    phi->set_ssa_temp_index(graph_->alloc_ssa_temp_index());
 | 
| -    phis_.Add(phi);  // Postpone phi insertion until after load forwarding.
 | 
| -
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      OS::Print("created pending phi %s for %s at B%" Pd "\n",
 | 
| -                phi->ToCString(),
 | 
| -                aliased_set_->places()[place_id]->ToCString(),
 | 
| -                block->block_id());
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Iterate over basic blocks and replace exposed loads with incoming
 | 
| -  // values.
 | 
| -  void ForwardLoads() {
 | 
| -    for (BlockIterator block_it = graph_->reverse_postorder_iterator();
 | 
| -         !block_it.Done();
 | 
| -         block_it.Advance()) {
 | 
| -      BlockEntryInstr* block = block_it.Current();
 | 
| -
 | 
| -      ZoneGrowableArray<Definition*>* loads =
 | 
| -          exposed_values_[block->preorder_number()];
 | 
| -      if (loads == NULL) continue;  // No exposed loads.
 | 
| -
 | 
| -      BitVector* in = in_[block->preorder_number()];
 | 
| -
 | 
| -      for (intptr_t i = 0; i < loads->length(); i++) {
 | 
| -        Definition* load = (*loads)[i];
 | 
| -        if (!in->Contains(load->place_id())) continue;  // No incoming value.
 | 
| -
 | 
| -        Definition* replacement = MergeIncomingValues(block, load->place_id());
 | 
| -        ASSERT(replacement != NULL);
 | 
| -
 | 
| -        // Sets of outgoing values are not linked into use lists so
 | 
| -        // they might contain values that were replace and removed
 | 
| -        // from the graph by this iteration.
 | 
| -        // To prevent using them we additionally mark definitions themselves
 | 
| -        // as replaced and store a pointer to the replacement.
 | 
| -        replacement = replacement->Replacement();
 | 
| -
 | 
| -        if (load != replacement) {
 | 
| -          EnsureSSATempIndex(graph_, load, replacement);
 | 
| -
 | 
| -          if (FLAG_trace_optimization) {
 | 
| -            OS::Print("Replacing load v%" Pd " with v%" Pd "\n",
 | 
| -                      load->ssa_temp_index(),
 | 
| -                      replacement->ssa_temp_index());
 | 
| -          }
 | 
| -
 | 
| -          load->ReplaceUsesWith(replacement);
 | 
| -          load->RemoveFromGraph();
 | 
| -          load->SetReplacement(replacement);
 | 
| -          forwarded_ = true;
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Check if the given phi take the same value on all code paths.
 | 
| -  // Eliminate it as redundant if this is the case.
 | 
| -  // When analyzing phi operands assumes that only generated during
 | 
| -  // this load phase can be redundant. They can be distinguished because
 | 
| -  // they are not marked alive.
 | 
| -  // TODO(vegorov): move this into a separate phase over all phis.
 | 
| -  bool EliminateRedundantPhi(PhiInstr* phi) {
 | 
| -    Definition* value = NULL;  // Possible value of this phi.
 | 
| -
 | 
| -    worklist_.Clear();
 | 
| -    if (in_worklist_ == NULL) {
 | 
| -      in_worklist_ = new(I) BitVector(I, graph_->current_ssa_temp_index());
 | 
| -    } else {
 | 
| -      in_worklist_->Clear();
 | 
| -    }
 | 
| -
 | 
| -    worklist_.Add(phi);
 | 
| -    in_worklist_->Add(phi->ssa_temp_index());
 | 
| -
 | 
| -    for (intptr_t i = 0; i < worklist_.length(); i++) {
 | 
| -      PhiInstr* phi = worklist_[i];
 | 
| -
 | 
| -      for (intptr_t i = 0; i < phi->InputCount(); i++) {
 | 
| -        Definition* input = phi->InputAt(i)->definition();
 | 
| -        if (input == phi) continue;
 | 
| -
 | 
| -        PhiInstr* phi_input = input->AsPhi();
 | 
| -        if ((phi_input != NULL) && !phi_input->is_alive()) {
 | 
| -          if (!in_worklist_->Contains(phi_input->ssa_temp_index())) {
 | 
| -            worklist_.Add(phi_input);
 | 
| -            in_worklist_->Add(phi_input->ssa_temp_index());
 | 
| -          }
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        if (value == NULL) {
 | 
| -          value = input;
 | 
| -        } else if (value != input) {
 | 
| -          return false;  // This phi is not redundant.
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    // All phis in the worklist are redundant and have the same computed
 | 
| -    // value on all code paths.
 | 
| -    ASSERT(value != NULL);
 | 
| -    for (intptr_t i = 0; i < worklist_.length(); i++) {
 | 
| -      worklist_[i]->ReplaceUsesWith(value);
 | 
| -    }
 | 
| -
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  // Returns true if definitions are congruent assuming their inputs
 | 
| -  // are congruent.
 | 
| -  bool CanBeCongruent(Definition* a, Definition* b) {
 | 
| -    return (a->tag() == b->tag()) &&
 | 
| -       ((a->IsPhi() && (a->GetBlock() == b->GetBlock())) ||
 | 
| -        (a->AllowsCSE() && a->Dependencies().IsNone() &&
 | 
| -         a->AttributesEqual(b)));
 | 
| -  }
 | 
| -
 | 
| -  // Given two definitions check if they are congruent under assumption that
 | 
| -  // their inputs will be proven congruent. If they are - add them to the
 | 
| -  // worklist to check their inputs' congruency.
 | 
| -  // Returns true if pair was added to the worklist or is already in the
 | 
| -  // worklist and false if a and b are not congruent.
 | 
| -  bool AddPairToCongruencyWorklist(Definition* a, Definition* b) {
 | 
| -    if (!CanBeCongruent(a, b)) {
 | 
| -      return false;
 | 
| -    }
 | 
| -
 | 
| -    // If a is already in the worklist check if it is being compared to b.
 | 
| -    // Give up if it is not.
 | 
| -    if (in_worklist_->Contains(a->ssa_temp_index())) {
 | 
| -      for (intptr_t i = 0; i < congruency_worklist_.length(); i += 2) {
 | 
| -        if (a == congruency_worklist_[i]) {
 | 
| -          return (b == congruency_worklist_[i + 1]);
 | 
| -        }
 | 
| -      }
 | 
| -      UNREACHABLE();
 | 
| -    } else if (in_worklist_->Contains(b->ssa_temp_index())) {
 | 
| -      return AddPairToCongruencyWorklist(b, a);
 | 
| -    }
 | 
| -
 | 
| -    congruency_worklist_.Add(a);
 | 
| -    congruency_worklist_.Add(b);
 | 
| -    in_worklist_->Add(a->ssa_temp_index());
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  bool AreInputsCongruent(Definition* a, Definition* b) {
 | 
| -    ASSERT(a->tag() == b->tag());
 | 
| -    ASSERT(a->InputCount() == b->InputCount());
 | 
| -    for (intptr_t j = 0; j < a->InputCount(); j++) {
 | 
| -      Definition* inputA = a->InputAt(j)->definition();
 | 
| -      Definition* inputB = b->InputAt(j)->definition();
 | 
| -
 | 
| -      if (inputA != inputB) {
 | 
| -        if (!AddPairToCongruencyWorklist(inputA, inputB)) {
 | 
| -          return false;
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  // Returns true if instruction dom dominates instruction other.
 | 
| -  static bool Dominates(Instruction* dom, Instruction* other) {
 | 
| -    BlockEntryInstr* dom_block = dom->GetBlock();
 | 
| -    BlockEntryInstr* other_block = other->GetBlock();
 | 
| -
 | 
| -    if (dom_block == other_block) {
 | 
| -      for (Instruction* current = dom->next();
 | 
| -           current != NULL;
 | 
| -           current = current->next()) {
 | 
| -        if (current == other) {
 | 
| -          return true;
 | 
| -        }
 | 
| -      }
 | 
| -      return false;
 | 
| -    }
 | 
| -
 | 
| -    return dom_block->Dominates(other_block);
 | 
| -  }
 | 
| -
 | 
| -  // Replace the given phi with another if they are congruent.
 | 
| -  // Returns true if succeeds.
 | 
| -  bool ReplacePhiWith(PhiInstr* phi, PhiInstr* replacement) {
 | 
| -    ASSERT(phi->InputCount() == replacement->InputCount());
 | 
| -    ASSERT(phi->block() == replacement->block());
 | 
| -
 | 
| -    congruency_worklist_.Clear();
 | 
| -    if (in_worklist_ == NULL) {
 | 
| -      in_worklist_ = new(I) BitVector(I, graph_->current_ssa_temp_index());
 | 
| -    } else {
 | 
| -      in_worklist_->Clear();
 | 
| -    }
 | 
| -
 | 
| -    // During the comparison worklist contains pairs of definitions to be
 | 
| -    // compared.
 | 
| -    if (!AddPairToCongruencyWorklist(phi, replacement)) {
 | 
| -      return false;
 | 
| -    }
 | 
| -
 | 
| -    // Process the worklist. It might grow during each comparison step.
 | 
| -    for (intptr_t i = 0; i < congruency_worklist_.length(); i += 2) {
 | 
| -      if (!AreInputsCongruent(congruency_worklist_[i],
 | 
| -                              congruency_worklist_[i + 1])) {
 | 
| -        return false;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    // At this point worklist contains pairs of congruent definitions.
 | 
| -    // Replace the one member of the pair with another maintaining proper
 | 
| -    // domination relation between definitions and uses.
 | 
| -    for (intptr_t i = 0; i < congruency_worklist_.length(); i += 2) {
 | 
| -      Definition* a = congruency_worklist_[i];
 | 
| -      Definition* b = congruency_worklist_[i + 1];
 | 
| -
 | 
| -      // If these definitions are not phis then we need to pick up one
 | 
| -      // that dominates another as the replacement: if a dominates b swap them.
 | 
| -      // Note: both a and b are used as a phi input at the same block B which
 | 
| -      // means a dominates B and b dominates B, which guarantees that either
 | 
| -      // a dominates b or b dominates a.
 | 
| -      if (!a->IsPhi()) {
 | 
| -        if (Dominates(a, b)) {
 | 
| -          Definition* t = a;
 | 
| -          a = b;
 | 
| -          b = t;
 | 
| -        }
 | 
| -        ASSERT(Dominates(b, a));
 | 
| -      }
 | 
| -
 | 
| -      if (FLAG_trace_load_optimization) {
 | 
| -        OS::Print("Replacing %s with congruent %s\n",
 | 
| -                  a->ToCString(),
 | 
| -                  b->ToCString());
 | 
| -      }
 | 
| -
 | 
| -      a->ReplaceUsesWith(b);
 | 
| -      if (a->IsPhi()) {
 | 
| -        // We might be replacing a phi introduced by the load forwarding
 | 
| -        // that is not inserted in the graph yet.
 | 
| -        ASSERT(b->IsPhi());
 | 
| -        PhiInstr* phi_a = a->AsPhi();
 | 
| -        if (phi_a->is_alive()) {
 | 
| -          phi_a->mark_dead();
 | 
| -          phi_a->block()->RemovePhi(phi_a);
 | 
| -          phi_a->UnuseAllInputs();
 | 
| -        }
 | 
| -      } else {
 | 
| -        a->RemoveFromGraph();
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  // Insert the given phi into the graph. Attempt to find an equal one in the
 | 
| -  // target block first.
 | 
| -  // Returns true if the phi was inserted and false if it was replaced.
 | 
| -  bool EmitPhi(PhiInstr* phi) {
 | 
| -    for (PhiIterator it(phi->block()); !it.Done(); it.Advance()) {
 | 
| -      if (ReplacePhiWith(phi, it.Current())) {
 | 
| -        return false;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    phi->mark_alive();
 | 
| -    phi->block()->InsertPhi(phi);
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  // Phis have not yet been inserted into the graph but they have uses of
 | 
| -  // their inputs.  Insert the non-redundant ones and clear the input uses
 | 
| -  // of the redundant ones.
 | 
| -  void EmitPhis() {
 | 
| -    // First eliminate all redundant phis.
 | 
| -    for (intptr_t i = 0; i < phis_.length(); i++) {
 | 
| -      PhiInstr* phi = phis_[i];
 | 
| -      if (!phi->HasUses() || EliminateRedundantPhi(phi)) {
 | 
| -        phi->UnuseAllInputs();
 | 
| -        phis_[i] = NULL;
 | 
| -      }
 | 
| -    }
 | 
| -
 | 
| -    // Now emit phis or replace them with equal phis already present in the
 | 
| -    // graph.
 | 
| -    for (intptr_t i = 0; i < phis_.length(); i++) {
 | 
| -      PhiInstr* phi = phis_[i];
 | 
| -      if ((phi != NULL) && (!phi->HasUses() || !EmitPhi(phi))) {
 | 
| -        phi->UnuseAllInputs();
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  ZoneGrowableArray<Definition*>* CreateBlockOutValues() {
 | 
| -    ZoneGrowableArray<Definition*>* out =
 | 
| -        new(I) ZoneGrowableArray<Definition*>(aliased_set_->max_place_id());
 | 
| -    for (intptr_t i = 0; i < aliased_set_->max_place_id(); i++) {
 | 
| -      out->Add(NULL);
 | 
| -    }
 | 
| -    return out;
 | 
| -  }
 | 
| -
 | 
| -  FlowGraph* graph_;
 | 
| -  DirectChainedHashMap<PointerKeyValueTrait<Place> >* map_;
 | 
| -
 | 
| -  // Mapping between field offsets in words and expression ids of loads from
 | 
| -  // that offset.
 | 
| -  AliasedSet* aliased_set_;
 | 
| -
 | 
| -  // Per block sets of expression ids for loads that are: incoming (available
 | 
| -  // on the entry), outgoing (available on the exit), generated and killed.
 | 
| -  GrowableArray<BitVector*> in_;
 | 
| -  GrowableArray<BitVector*> out_;
 | 
| -  GrowableArray<BitVector*> gen_;
 | 
| -  GrowableArray<BitVector*> kill_;
 | 
| -
 | 
| -  // Per block list of upwards exposed loads.
 | 
| -  GrowableArray<ZoneGrowableArray<Definition*>*> exposed_values_;
 | 
| -
 | 
| -  // Per block mappings between expression ids and outgoing definitions that
 | 
| -  // represent those ids.
 | 
| -  GrowableArray<ZoneGrowableArray<Definition*>*> out_values_;
 | 
| -
 | 
| -  // List of phis generated during ComputeOutValues and ForwardLoads.
 | 
| -  // Some of these phis might be redundant and thus a separate pass is
 | 
| -  // needed to emit only non-redundant ones.
 | 
| -  GrowableArray<PhiInstr*> phis_;
 | 
| -
 | 
| -  // Auxiliary worklist used by redundant phi elimination.
 | 
| -  GrowableArray<PhiInstr*> worklist_;
 | 
| -  GrowableArray<Definition*> congruency_worklist_;
 | 
| -  BitVector* in_worklist_;
 | 
| -
 | 
| -
 | 
| -  // True if any load was eliminated.
 | 
| -  bool forwarded_;
 | 
| -
 | 
| -  DISALLOW_COPY_AND_ASSIGN(LoadOptimizer);
 | 
| -};
 | 
| -
 | 
| -
 | 
| -class StoreOptimizer : public LivenessAnalysis {
 | 
| - public:
 | 
| -  StoreOptimizer(FlowGraph* graph,
 | 
| -                 AliasedSet* aliased_set,
 | 
| -                 DirectChainedHashMap<PointerKeyValueTrait<Place> >* map)
 | 
| -      : LivenessAnalysis(aliased_set->max_place_id(), graph->postorder()),
 | 
| -        graph_(graph),
 | 
| -        map_(map),
 | 
| -        aliased_set_(aliased_set),
 | 
| -        exposed_stores_(graph_->postorder().length()) {
 | 
| -    const intptr_t num_blocks = graph_->postorder().length();
 | 
| -    for (intptr_t i = 0; i < num_blocks; i++) {
 | 
| -      exposed_stores_.Add(NULL);
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  static void OptimizeGraph(FlowGraph* graph) {
 | 
| -    ASSERT(FLAG_load_cse);
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      FlowGraphPrinter::PrintGraph("Before StoreOptimizer", graph);
 | 
| -    }
 | 
| -
 | 
| -    DirectChainedHashMap<PointerKeyValueTrait<Place> > map;
 | 
| -    AliasedSet* aliased_set = NumberPlaces(graph, &map, kOptimizeStores);
 | 
| -    if ((aliased_set != NULL) && !aliased_set->IsEmpty()) {
 | 
| -      StoreOptimizer store_optimizer(graph, aliased_set, &map);
 | 
| -      store_optimizer.Optimize();
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| - private:
 | 
| -  void Optimize() {
 | 
| -    Analyze();
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      Dump();
 | 
| -    }
 | 
| -    EliminateDeadStores();
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      FlowGraphPrinter::PrintGraph("After StoreOptimizer", graph_);
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  bool CanEliminateStore(Instruction* instr) {
 | 
| -    switch (instr->tag()) {
 | 
| -      case Instruction::kStoreInstanceField:
 | 
| -        if (instr->AsStoreInstanceField()->is_initialization()) {
 | 
| -          // Can't eliminate stores that initialized unboxed fields.
 | 
| -          return false;
 | 
| -        }
 | 
| -      case Instruction::kStoreIndexed:
 | 
| -      case Instruction::kStoreStaticField:
 | 
| -        return true;
 | 
| -      default:
 | 
| -        UNREACHABLE();
 | 
| -        return false;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  virtual void ComputeInitialSets() {
 | 
| -    Isolate* isolate = graph_->isolate();
 | 
| -    BitVector* all_places = new(isolate) BitVector(isolate,
 | 
| -        aliased_set_->max_place_id());
 | 
| -    all_places->SetAll();
 | 
| -    for (BlockIterator block_it = graph_->postorder_iterator();
 | 
| -         !block_it.Done();
 | 
| -         block_it.Advance()) {
 | 
| -      BlockEntryInstr* block = block_it.Current();
 | 
| -      const intptr_t postorder_number = block->postorder_number();
 | 
| -
 | 
| -      BitVector* kill = kill_[postorder_number];
 | 
| -      BitVector* live_in = live_in_[postorder_number];
 | 
| -      BitVector* live_out = live_out_[postorder_number];
 | 
| -
 | 
| -      ZoneGrowableArray<Instruction*>* exposed_stores = NULL;
 | 
| -
 | 
| -      // Iterate backwards starting at the last instruction.
 | 
| -      for (BackwardInstructionIterator instr_it(block);
 | 
| -           !instr_it.Done();
 | 
| -           instr_it.Advance()) {
 | 
| -        Instruction* instr = instr_it.Current();
 | 
| -
 | 
| -        bool is_load = false;
 | 
| -        bool is_store = false;
 | 
| -        Place place(instr, &is_load, &is_store);
 | 
| -        if (place.IsFinalField()) {
 | 
| -          // Loads/stores of final fields do not participate.
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        // Handle stores.
 | 
| -        if (is_store) {
 | 
| -          if (kill->Contains(instr->place_id())) {
 | 
| -            if (!live_in->Contains(instr->place_id()) &&
 | 
| -                CanEliminateStore(instr)) {
 | 
| -              if (FLAG_trace_optimization) {
 | 
| -                OS::Print(
 | 
| -                    "Removing dead store to place %" Pd " in block B%" Pd "\n",
 | 
| -                    instr->place_id(), block->block_id());
 | 
| -              }
 | 
| -              instr_it.RemoveCurrentFromGraph();
 | 
| -            }
 | 
| -          } else if (!live_in->Contains(instr->place_id())) {
 | 
| -            // Mark this store as down-ward exposed: They are the only
 | 
| -            // candidates for the global store elimination.
 | 
| -            if (exposed_stores == NULL) {
 | 
| -              const intptr_t kMaxExposedStoresInitialSize = 5;
 | 
| -              exposed_stores = new(isolate) ZoneGrowableArray<Instruction*>(
 | 
| -                  Utils::Minimum(kMaxExposedStoresInitialSize,
 | 
| -                                 aliased_set_->max_place_id()));
 | 
| -            }
 | 
| -            exposed_stores->Add(instr);
 | 
| -          }
 | 
| -          // Interfering stores kill only loads from the same place.
 | 
| -          kill->Add(instr->place_id());
 | 
| -          live_in->Remove(instr->place_id());
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        // Handle side effects, deoptimization and function return.
 | 
| -        if (!instr->Effects().IsNone() ||
 | 
| -            instr->CanDeoptimize() ||
 | 
| -            instr->IsThrow() ||
 | 
| -            instr->IsReThrow() ||
 | 
| -            instr->IsReturn()) {
 | 
| -          // Instructions that return from the function, instructions with side
 | 
| -          // effects and instructions that can deoptimize are considered as
 | 
| -          // loads from all places.
 | 
| -          live_in->CopyFrom(all_places);
 | 
| -          if (instr->IsThrow() || instr->IsReThrow() || instr->IsReturn()) {
 | 
| -            // Initialize live-out for exit blocks since it won't be computed
 | 
| -            // otherwise during the fixed point iteration.
 | 
| -            live_out->CopyFrom(all_places);
 | 
| -          }
 | 
| -          continue;
 | 
| -        }
 | 
| -
 | 
| -        // Handle loads.
 | 
| -        Definition* defn = instr->AsDefinition();
 | 
| -        if ((defn != NULL) && IsLoadEliminationCandidate(defn)) {
 | 
| -          const intptr_t alias = aliased_set_->LookupAliasId(place.ToAlias());
 | 
| -          live_in->AddAll(aliased_set_->GetKilledSet(alias));
 | 
| -          continue;
 | 
| -        }
 | 
| -      }
 | 
| -      exposed_stores_[postorder_number] = exposed_stores;
 | 
| -    }
 | 
| -    if (FLAG_trace_load_optimization) {
 | 
| -      Dump();
 | 
| -      OS::Print("---\n");
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  void EliminateDeadStores() {
 | 
| -    // Iteration order does not matter here.
 | 
| -    for (BlockIterator block_it = graph_->postorder_iterator();
 | 
| -         !block_it.Done();
 | 
| -         block_it.Advance()) {
 | 
| -      BlockEntryInstr* block = block_it.Current();
 | 
| -      const intptr_t postorder_number = block->postorder_number();
 | 
| -
 | 
| -      BitVector* live_out = live_out_[postorder_number];
 | 
| -
 | 
| -      ZoneGrowableArray<Instruction*>* exposed_stores =
 | 
| -        exposed_stores_[postorder_number];
 | 
| -      if (exposed_stores == NULL) continue;  // No exposed stores.
 | 
| -
 | 
| -      // Iterate over candidate stores.
 | 
| -      for (intptr_t i = 0; i < exposed_stores->length(); ++i) {
 | 
| -        Instruction* instr = (*exposed_stores)[i];
 | 
| -        bool is_load = false;
 | 
| -        bool is_store = false;
 | 
| -        Place place(instr, &is_load, &is_store);
 | 
| -        ASSERT(!is_load && is_store);
 | 
| -        if (place.IsFinalField()) {
 | 
| -          // Final field do not participate in dead store elimination.
 | 
| -          continue;
 | 
| -        }
 | 
| -        // Eliminate a downward exposed store if the corresponding place is not
 | 
| -        // in live-out.
 | 
| -        if (!live_out->Contains(instr->place_id()) &&
 | 
| -            CanEliminateStore(instr)) {
 | 
| -          if (FLAG_trace_optimization) {
 | 
| -            OS::Print("Removing dead store to place %" Pd " block B%" Pd "\n",
 | 
| -                      instr->place_id(), block->block_id());
 | 
| -          }
 | 
| -          instr->RemoveFromGraph(/* ignored */ false);
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  FlowGraph* graph_;
 | 
| -  DirectChainedHashMap<PointerKeyValueTrait<Place> >* map_;
 | 
| -
 | 
| -  // Mapping between field offsets in words and expression ids of loads from
 | 
| -  // that offset.
 | 
| -  AliasedSet* aliased_set_;
 | 
| -
 | 
| -  // Per block list of downward exposed stores.
 | 
| -  GrowableArray<ZoneGrowableArray<Instruction*>*> exposed_stores_;
 | 
| -
 | 
| -  DISALLOW_COPY_AND_ASSIGN(StoreOptimizer);
 | 
| -};
 | 
| -
 | 
| -
 | 
| -void DeadStoreElimination::Optimize(FlowGraph* graph) {
 | 
| -  if (FLAG_dead_store_elimination) {
 | 
| -    StoreOptimizer::OptimizeGraph(graph);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Returns true iff this definition is used in a non-phi instruction.
 | 
| -static bool HasRealUse(Definition* def) {
 | 
| -  // Environment uses are real (non-phi) uses.
 | 
| -  if (def->env_use_list() != NULL) return true;
 | 
| -
 | 
| -  for (Value::Iterator it(def->input_use_list());
 | 
| -       !it.Done();
 | 
| -       it.Advance()) {
 | 
| -    if (!it.Current()->instruction()->IsPhi()) return true;
 | 
| -  }
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void DeadCodeElimination::EliminateDeadPhis(FlowGraph* flow_graph) {
 | 
| -  GrowableArray<PhiInstr*> live_phis;
 | 
| -  for (BlockIterator b = flow_graph->postorder_iterator();
 | 
| -       !b.Done();
 | 
| -       b.Advance()) {
 | 
| -    JoinEntryInstr* join = b.Current()->AsJoinEntry();
 | 
| -    if (join != NULL) {
 | 
| -      for (PhiIterator it(join); !it.Done(); it.Advance()) {
 | 
| -        PhiInstr* phi = it.Current();
 | 
| -        // Phis that have uses and phis inside try blocks are
 | 
| -        // marked as live.
 | 
| -        if (HasRealUse(phi) || join->InsideTryBlock()) {
 | 
| -          live_phis.Add(phi);
 | 
| -          phi->mark_alive();
 | 
| -        } else {
 | 
| -          phi->mark_dead();
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  while (!live_phis.is_empty()) {
 | 
| -    PhiInstr* phi = live_phis.RemoveLast();
 | 
| -    for (intptr_t i = 0; i < phi->InputCount(); i++) {
 | 
| -      Value* val = phi->InputAt(i);
 | 
| -      PhiInstr* used_phi = val->definition()->AsPhi();
 | 
| -      if ((used_phi != NULL) && !used_phi->is_alive()) {
 | 
| -        used_phi->mark_alive();
 | 
| -        live_phis.Add(used_phi);
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  for (BlockIterator it(flow_graph->postorder_iterator());
 | 
| -       !it.Done();
 | 
| -       it.Advance()) {
 | 
| -    JoinEntryInstr* join = it.Current()->AsJoinEntry();
 | 
| -    if (join != NULL) {
 | 
| -      if (join->phis_ == NULL) continue;
 | 
| -
 | 
| -      // Eliminate dead phis and compact the phis_ array of the block.
 | 
| -      intptr_t to_index = 0;
 | 
| -      for (intptr_t i = 0; i < join->phis_->length(); ++i) {
 | 
| -        PhiInstr* phi = (*join->phis_)[i];
 | 
| -        if (phi != NULL) {
 | 
| -          if (!phi->is_alive()) {
 | 
| -            phi->ReplaceUsesWith(flow_graph->constant_null());
 | 
| -            phi->UnuseAllInputs();
 | 
| -            (*join->phis_)[i] = NULL;
 | 
| -            if (FLAG_trace_optimization) {
 | 
| -              OS::Print("Removing dead phi v%" Pd "\n", phi->ssa_temp_index());
 | 
| -            }
 | 
| -          } else if (phi->IsRedundant()) {
 | 
| -            phi->ReplaceUsesWith(phi->InputAt(0)->definition());
 | 
| -            phi->UnuseAllInputs();
 | 
| -            (*join->phis_)[i] = NULL;
 | 
| -            if (FLAG_trace_optimization) {
 | 
| -              OS::Print("Removing redundant phi v%" Pd "\n",
 | 
| -                         phi->ssa_temp_index());
 | 
| -            }
 | 
| -          } else {
 | 
| -            (*join->phis_)[to_index++] = phi;
 | 
| -          }
 | 
| -        }
 | 
| -      }
 | 
| -      if (to_index == 0) {
 | 
| -        join->phis_ = NULL;
 | 
| -      } else {
 | 
| -        join->phis_->TruncateTo(to_index);
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -class CSEInstructionMap : public ValueObject {
 | 
| - public:
 | 
| -  // Right now CSE and LICM track a single effect: possible externalization of
 | 
| -  // strings.
 | 
| -  // Other effects like modifications of fields are tracked in a separate load
 | 
| -  // forwarding pass via Alias structure.
 | 
| -  COMPILE_ASSERT(EffectSet::kLastEffect == 1);
 | 
| -
 | 
| -  CSEInstructionMap() : independent_(), dependent_() { }
 | 
| -  explicit CSEInstructionMap(const CSEInstructionMap& other)
 | 
| -      : ValueObject(),
 | 
| -        independent_(other.independent_),
 | 
| -        dependent_(other.dependent_) {
 | 
| -  }
 | 
| -
 | 
| -  void RemoveAffected(EffectSet effects) {
 | 
| -    if (!effects.IsNone()) {
 | 
| -      dependent_.Clear();
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  Instruction* Lookup(Instruction* other) const {
 | 
| -    return GetMapFor(other)->Lookup(other);
 | 
| -  }
 | 
| -
 | 
| -  void Insert(Instruction* instr) {
 | 
| -    return GetMapFor(instr)->Insert(instr);
 | 
| -  }
 | 
| -
 | 
| - private:
 | 
| -  typedef DirectChainedHashMap<PointerKeyValueTrait<Instruction> >  Map;
 | 
| -
 | 
| -  Map* GetMapFor(Instruction* instr) {
 | 
| -    return instr->Dependencies().IsNone() ? &independent_ : &dependent_;
 | 
| -  }
 | 
| -
 | 
| -  const Map* GetMapFor(Instruction* instr) const {
 | 
| -    return instr->Dependencies().IsNone() ? &independent_ : &dependent_;
 | 
| -  }
 | 
| -
 | 
| -  // All computations that are not affected by any side-effect.
 | 
| -  // Majority of computations are not affected by anything and will be in
 | 
| -  // this map.
 | 
| -  Map independent_;
 | 
| -
 | 
| -  // All computations that are affected by side effect.
 | 
| -  Map dependent_;
 | 
| -};
 | 
| -
 | 
| -
 | 
| -bool DominatorBasedCSE::Optimize(FlowGraph* graph) {
 | 
| -  bool changed = false;
 | 
| -  if (FLAG_load_cse) {
 | 
| -    changed = LoadOptimizer::OptimizeGraph(graph) || changed;
 | 
| -  }
 | 
| -
 | 
| -  CSEInstructionMap map;
 | 
| -  changed = OptimizeRecursive(graph, graph->graph_entry(), &map) || changed;
 | 
| -
 | 
| -  return changed;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool DominatorBasedCSE::OptimizeRecursive(
 | 
| -    FlowGraph* graph,
 | 
| -    BlockEntryInstr* block,
 | 
| -    CSEInstructionMap* map) {
 | 
| -  bool changed = false;
 | 
| -  for (ForwardInstructionIterator it(block); !it.Done(); it.Advance()) {
 | 
| -    Instruction* current = it.Current();
 | 
| -    if (current->AllowsCSE()) {
 | 
| -      Instruction* replacement = map->Lookup(current);
 | 
| -      if ((replacement != NULL) &&
 | 
| -          graph->block_effects()->IsAvailableAt(replacement, block)) {
 | 
| -        // Replace current with lookup result.
 | 
| -        ReplaceCurrentInstruction(&it, current, replacement, graph);
 | 
| -        changed = true;
 | 
| -        continue;
 | 
| -      }
 | 
| -
 | 
| -      // For simplicity we assume that instruction either does not depend on
 | 
| -      // anything or does not affect anything. If this is not the case then
 | 
| -      // we should first remove affected instructions from the map and
 | 
| -      // then add instruction to the map so that it does not kill itself.
 | 
| -      ASSERT(current->Effects().IsNone() || current->Dependencies().IsNone());
 | 
| -      map->Insert(current);
 | 
| -    }
 | 
| -
 | 
| -    map->RemoveAffected(current->Effects());
 | 
| -  }
 | 
| -
 | 
| -  // Process children in the dominator tree recursively.
 | 
| -  intptr_t num_children = block->dominated_blocks().length();
 | 
| -  for (intptr_t i = 0; i < num_children; ++i) {
 | 
| -    BlockEntryInstr* child = block->dominated_blocks()[i];
 | 
| -    if (i  < num_children - 1) {
 | 
| -      // Copy map.
 | 
| -      CSEInstructionMap child_map(*map);
 | 
| -      changed = OptimizeRecursive(graph, child, &child_map) || changed;
 | 
| -    } else {
 | 
| -      // Reuse map for the last child.
 | 
| -      changed = OptimizeRecursive(graph, child, map) || changed;
 | 
| -    }
 | 
| -  }
 | 
| -  return changed;
 | 
| -}
 | 
| -
 | 
| -
 | 
|  ConstantPropagator::ConstantPropagator(
 | 
|      FlowGraph* graph,
 | 
|      const GrowableArray<BlockEntryInstr*>& ignored)
 | 
| @@ -8202,7 +725,7 @@
 | 
|      if ((checked_type.IsFloat32x4Type() && (rep == kUnboxedFloat32x4)) ||
 | 
|          (checked_type.IsInt32x4Type() && (rep == kUnboxedInt32x4)) ||
 | 
|          (checked_type.IsDoubleType() && (rep == kUnboxedDouble) &&
 | 
| -         CanUnboxDouble()) ||
 | 
| +         FlowGraphCompiler::SupportsUnboxedDoubles()) ||
 | 
|          (checked_type.IsIntType() && (rep == kUnboxedMint))) {
 | 
|        // Ensure that compile time type matches representation.
 | 
|        ASSERT(((rep == kUnboxedFloat32x4) && (value_cid == kFloat32x4Cid)) ||
 | 
| @@ -9112,976 +1635,4 @@
 | 
|    }
 | 
|  }
 | 
|  
 | 
| -
 | 
| -// Returns true if the given phi has a single input use and
 | 
| -// is used in the environments either at the corresponding block entry or
 | 
| -// at the same instruction where input use is.
 | 
| -static bool PhiHasSingleUse(PhiInstr* phi, Value* use) {
 | 
| -  if ((use->next_use() != NULL) || (phi->input_use_list() != use)) {
 | 
| -    return false;
 | 
| -  }
 | 
| -
 | 
| -  BlockEntryInstr* block = phi->block();
 | 
| -  for (Value* env_use = phi->env_use_list();
 | 
| -       env_use != NULL;
 | 
| -       env_use = env_use->next_use()) {
 | 
| -    if ((env_use->instruction() != block) &&
 | 
| -        (env_use->instruction() != use->instruction())) {
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -bool BranchSimplifier::Match(JoinEntryInstr* block) {
 | 
| -  // Match the pattern of a branch on a comparison whose left operand is a
 | 
| -  // phi from the same block, and whose right operand is a constant.
 | 
| -  //
 | 
| -  //   Branch(Comparison(kind, Phi, Constant))
 | 
| -  //
 | 
| -  // These are the branches produced by inlining in a test context.  Also,
 | 
| -  // the phi has no other uses so they can simply be eliminated.  The block
 | 
| -  // has no other phis and no instructions intervening between the phi and
 | 
| -  // branch so the block can simply be eliminated.
 | 
| -  BranchInstr* branch = block->last_instruction()->AsBranch();
 | 
| -  ASSERT(branch != NULL);
 | 
| -  ComparisonInstr* comparison = branch->comparison();
 | 
| -  Value* left = comparison->left();
 | 
| -  PhiInstr* phi = left->definition()->AsPhi();
 | 
| -  Value* right = comparison->right();
 | 
| -  ConstantInstr* constant =
 | 
| -      (right == NULL) ? NULL : right->definition()->AsConstant();
 | 
| -  return (phi != NULL) &&
 | 
| -      (constant != NULL) &&
 | 
| -      (phi->GetBlock() == block) &&
 | 
| -      PhiHasSingleUse(phi, left) &&
 | 
| -      (block->next() == branch) &&
 | 
| -      (block->phis()->length() == 1);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -JoinEntryInstr* BranchSimplifier::ToJoinEntry(Isolate* isolate,
 | 
| -                                              TargetEntryInstr* target) {
 | 
| -  // Convert a target block into a join block.  Branches will be duplicated
 | 
| -  // so the former true and false targets become joins of the control flows
 | 
| -  // from all the duplicated branches.
 | 
| -  JoinEntryInstr* join =
 | 
| -      new(isolate) JoinEntryInstr(target->block_id(), target->try_index());
 | 
| -  join->InheritDeoptTarget(isolate, target);
 | 
| -  join->LinkTo(target->next());
 | 
| -  join->set_last_instruction(target->last_instruction());
 | 
| -  target->UnuseAllInputs();
 | 
| -  return join;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -BranchInstr* BranchSimplifier::CloneBranch(Isolate* isolate,
 | 
| -                                           BranchInstr* branch,
 | 
| -                                           Value* new_left,
 | 
| -                                           Value* new_right) {
 | 
| -  ComparisonInstr* comparison = branch->comparison();
 | 
| -  ComparisonInstr* new_comparison =
 | 
| -      comparison->CopyWithNewOperands(new_left, new_right);
 | 
| -  BranchInstr* new_branch = new(isolate) BranchInstr(new_comparison);
 | 
| -  new_branch->set_is_checked(branch->is_checked());
 | 
| -  return new_branch;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void BranchSimplifier::Simplify(FlowGraph* flow_graph) {
 | 
| -  // Optimize some branches that test the value of a phi.  When it is safe
 | 
| -  // to do so, push the branch to each of the predecessor blocks.  This is
 | 
| -  // an optimization when (a) it can avoid materializing a boolean object at
 | 
| -  // the phi only to test its value, and (b) it can expose opportunities for
 | 
| -  // constant propagation and unreachable code elimination.  This
 | 
| -  // optimization is intended to run after inlining which creates
 | 
| -  // opportunities for optimization (a) and before constant folding which
 | 
| -  // can perform optimization (b).
 | 
| -
 | 
| -  // Begin with a worklist of join blocks ending in branches.  They are
 | 
| -  // candidates for the pattern below.
 | 
| -  Isolate* isolate = flow_graph->isolate();
 | 
| -  const GrowableArray<BlockEntryInstr*>& postorder = flow_graph->postorder();
 | 
| -  GrowableArray<BlockEntryInstr*> worklist(postorder.length());
 | 
| -  for (BlockIterator it(postorder); !it.Done(); it.Advance()) {
 | 
| -    BlockEntryInstr* block = it.Current();
 | 
| -    if (block->IsJoinEntry() && block->last_instruction()->IsBranch()) {
 | 
| -      worklist.Add(block);
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Rewrite until no more instance of the pattern exists.
 | 
| -  bool changed = false;
 | 
| -  while (!worklist.is_empty()) {
 | 
| -    // All blocks in the worklist are join blocks (ending with a branch).
 | 
| -    JoinEntryInstr* block = worklist.RemoveLast()->AsJoinEntry();
 | 
| -    ASSERT(block != NULL);
 | 
| -
 | 
| -    if (Match(block)) {
 | 
| -      changed = true;
 | 
| -
 | 
| -      // The branch will be copied and pushed to all the join's
 | 
| -      // predecessors.  Convert the true and false target blocks into join
 | 
| -      // blocks to join the control flows from all of the true
 | 
| -      // (respectively, false) targets of the copied branches.
 | 
| -      //
 | 
| -      // The converted join block will have no phis, so it cannot be another
 | 
| -      // instance of the pattern.  There is thus no need to add it to the
 | 
| -      // worklist.
 | 
| -      BranchInstr* branch = block->last_instruction()->AsBranch();
 | 
| -      ASSERT(branch != NULL);
 | 
| -      JoinEntryInstr* join_true =
 | 
| -          ToJoinEntry(isolate, branch->true_successor());
 | 
| -      JoinEntryInstr* join_false =
 | 
| -          ToJoinEntry(isolate, branch->false_successor());
 | 
| -
 | 
| -      ComparisonInstr* comparison = branch->comparison();
 | 
| -      PhiInstr* phi = comparison->left()->definition()->AsPhi();
 | 
| -      ConstantInstr* constant = comparison->right()->definition()->AsConstant();
 | 
| -      ASSERT(constant != NULL);
 | 
| -      // Copy the constant and branch and push it to all the predecessors.
 | 
| -      for (intptr_t i = 0, count = block->PredecessorCount(); i < count; ++i) {
 | 
| -        GotoInstr* old_goto =
 | 
| -            block->PredecessorAt(i)->last_instruction()->AsGoto();
 | 
| -        ASSERT(old_goto != NULL);
 | 
| -
 | 
| -        // Replace the goto in each predecessor with a rewritten branch,
 | 
| -        // rewritten to use the corresponding phi input instead of the phi.
 | 
| -        Value* new_left = phi->InputAt(i)->Copy(isolate);
 | 
| -        Value* new_right = new(isolate) Value(constant);
 | 
| -        BranchInstr* new_branch =
 | 
| -            CloneBranch(isolate, branch, new_left, new_right);
 | 
| -        if (branch->env() == NULL) {
 | 
| -          new_branch->InheritDeoptTarget(isolate, old_goto);
 | 
| -        } else {
 | 
| -          // Take the environment from the branch if it has one.
 | 
| -          new_branch->InheritDeoptTarget(isolate, branch);
 | 
| -          // InheritDeoptTarget gave the new branch's comparison the same
 | 
| -          // deopt id that it gave the new branch.  The id should be the
 | 
| -          // deopt id of the original comparison.
 | 
| -          new_branch->comparison()->SetDeoptId(*comparison);
 | 
| -          // The phi can be used in the branch's environment.  Rename such
 | 
| -          // uses.
 | 
| -          for (Environment::DeepIterator it(new_branch->env());
 | 
| -               !it.Done();
 | 
| -               it.Advance()) {
 | 
| -            Value* use = it.CurrentValue();
 | 
| -            if (use->definition() == phi) {
 | 
| -              Definition* replacement = phi->InputAt(i)->definition();
 | 
| -              use->RemoveFromUseList();
 | 
| -              use->set_definition(replacement);
 | 
| -              replacement->AddEnvUse(use);
 | 
| -            }
 | 
| -          }
 | 
| -        }
 | 
| -
 | 
| -        new_branch->InsertBefore(old_goto);
 | 
| -        new_branch->set_next(NULL);  // Detaching the goto from the graph.
 | 
| -        old_goto->UnuseAllInputs();
 | 
| -
 | 
| -        // Update the predecessor block.  We may have created another
 | 
| -        // instance of the pattern so add it to the worklist if necessary.
 | 
| -        BlockEntryInstr* branch_block = new_branch->GetBlock();
 | 
| -        branch_block->set_last_instruction(new_branch);
 | 
| -        if (branch_block->IsJoinEntry()) worklist.Add(branch_block);
 | 
| -
 | 
| -        // Connect the branch to the true and false joins, via empty target
 | 
| -        // blocks.
 | 
| -        TargetEntryInstr* true_target =
 | 
| -            new(isolate) TargetEntryInstr(flow_graph->max_block_id() + 1,
 | 
| -                                          block->try_index());
 | 
| -        true_target->InheritDeoptTarget(isolate, join_true);
 | 
| -        TargetEntryInstr* false_target =
 | 
| -            new(isolate) TargetEntryInstr(flow_graph->max_block_id() + 2,
 | 
| -                                          block->try_index());
 | 
| -        false_target->InheritDeoptTarget(isolate, join_false);
 | 
| -        flow_graph->set_max_block_id(flow_graph->max_block_id() + 2);
 | 
| -        *new_branch->true_successor_address() = true_target;
 | 
| -        *new_branch->false_successor_address() = false_target;
 | 
| -        GotoInstr* goto_true = new(isolate) GotoInstr(join_true);
 | 
| -        goto_true->InheritDeoptTarget(isolate, join_true);
 | 
| -        true_target->LinkTo(goto_true);
 | 
| -        true_target->set_last_instruction(goto_true);
 | 
| -        GotoInstr* goto_false = new(isolate) GotoInstr(join_false);
 | 
| -        goto_false->InheritDeoptTarget(isolate, join_false);
 | 
| -        false_target->LinkTo(goto_false);
 | 
| -        false_target->set_last_instruction(goto_false);
 | 
| -      }
 | 
| -      // When all predecessors have been rewritten, the original block is
 | 
| -      // unreachable from the graph.
 | 
| -      phi->UnuseAllInputs();
 | 
| -      branch->UnuseAllInputs();
 | 
| -      block->UnuseAllInputs();
 | 
| -      ASSERT(!phi->HasUses());
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (changed) {
 | 
| -    // We may have changed the block order and the dominator tree.
 | 
| -    flow_graph->DiscoverBlocks();
 | 
| -    GrowableArray<BitVector*> dominance_frontier;
 | 
| -    flow_graph->ComputeDominators(&dominance_frontier);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static bool IsTrivialBlock(BlockEntryInstr* block, Definition* defn) {
 | 
| -  return (block->IsTargetEntry() && (block->PredecessorCount() == 1)) &&
 | 
| -    ((block->next() == block->last_instruction()) ||
 | 
| -     ((block->next() == defn) && (defn->next() == block->last_instruction())));
 | 
| -}
 | 
| -
 | 
| -
 | 
| -static void EliminateTrivialBlock(BlockEntryInstr* block,
 | 
| -                                  Definition* instr,
 | 
| -                                  IfThenElseInstr* before) {
 | 
| -  block->UnuseAllInputs();
 | 
| -  block->last_instruction()->UnuseAllInputs();
 | 
| -
 | 
| -  if ((block->next() == instr) &&
 | 
| -      (instr->next() == block->last_instruction())) {
 | 
| -    before->previous()->LinkTo(instr);
 | 
| -    instr->LinkTo(before);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void IfConverter::Simplify(FlowGraph* flow_graph) {
 | 
| -  Isolate* isolate = flow_graph->isolate();
 | 
| -  bool changed = false;
 | 
| -
 | 
| -  const GrowableArray<BlockEntryInstr*>& postorder = flow_graph->postorder();
 | 
| -  for (BlockIterator it(postorder); !it.Done(); it.Advance()) {
 | 
| -    BlockEntryInstr* block = it.Current();
 | 
| -    JoinEntryInstr* join = block->AsJoinEntry();
 | 
| -
 | 
| -    // Detect diamond control flow pattern which materializes a value depending
 | 
| -    // on the result of the comparison:
 | 
| -    //
 | 
| -    // B_pred:
 | 
| -    //   ...
 | 
| -    //   Branch if COMP goto (B_pred1, B_pred2)
 | 
| -    // B_pred1: -- trivial block that contains at most one definition
 | 
| -    //   v1 = Constant(...)
 | 
| -    //   goto B_block
 | 
| -    // B_pred2: -- trivial block that contains at most one definition
 | 
| -    //   v2 = Constant(...)
 | 
| -    //   goto B_block
 | 
| -    // B_block:
 | 
| -    //   v3 = phi(v1, v2) -- single phi
 | 
| -    //
 | 
| -    // and replace it with
 | 
| -    //
 | 
| -    // Ba:
 | 
| -    //   v3 = IfThenElse(COMP ? v1 : v2)
 | 
| -    //
 | 
| -    if ((join != NULL) &&
 | 
| -        (join->phis() != NULL) &&
 | 
| -        (join->phis()->length() == 1) &&
 | 
| -        (block->PredecessorCount() == 2)) {
 | 
| -      BlockEntryInstr* pred1 = block->PredecessorAt(0);
 | 
| -      BlockEntryInstr* pred2 = block->PredecessorAt(1);
 | 
| -
 | 
| -      PhiInstr* phi = (*join->phis())[0];
 | 
| -      Value* v1 = phi->InputAt(0);
 | 
| -      Value* v2 = phi->InputAt(1);
 | 
| -
 | 
| -      if (IsTrivialBlock(pred1, v1->definition()) &&
 | 
| -          IsTrivialBlock(pred2, v2->definition()) &&
 | 
| -          (pred1->PredecessorAt(0) == pred2->PredecessorAt(0))) {
 | 
| -        BlockEntryInstr* pred = pred1->PredecessorAt(0);
 | 
| -        BranchInstr* branch = pred->last_instruction()->AsBranch();
 | 
| -        ComparisonInstr* comparison = branch->comparison();
 | 
| -
 | 
| -        // Check if the platform supports efficient branchless IfThenElseInstr
 | 
| -        // for the given combination of comparison and values flowing from
 | 
| -        // false and true paths.
 | 
| -        if (IfThenElseInstr::Supports(comparison, v1, v2)) {
 | 
| -          Value* if_true = (pred1 == branch->true_successor()) ? v1 : v2;
 | 
| -          Value* if_false = (pred2 == branch->true_successor()) ? v1 : v2;
 | 
| -
 | 
| -          ComparisonInstr* new_comparison =
 | 
| -              comparison->CopyWithNewOperands(
 | 
| -                  comparison->left()->Copy(isolate),
 | 
| -                  comparison->right()->Copy(isolate));
 | 
| -          IfThenElseInstr* if_then_else = new(isolate) IfThenElseInstr(
 | 
| -              new_comparison,
 | 
| -              if_true->Copy(isolate),
 | 
| -              if_false->Copy(isolate));
 | 
| -          flow_graph->InsertBefore(branch,
 | 
| -                                   if_then_else,
 | 
| -                                   NULL,
 | 
| -                                   FlowGraph::kValue);
 | 
| -
 | 
| -          phi->ReplaceUsesWith(if_then_else);
 | 
| -
 | 
| -          // Connect IfThenElseInstr to the first instruction in the merge block
 | 
| -          // effectively eliminating diamond control flow.
 | 
| -          // Current block as well as pred1 and pred2 blocks are no longer in
 | 
| -          // the graph at this point.
 | 
| -          if_then_else->LinkTo(join->next());
 | 
| -          pred->set_last_instruction(join->last_instruction());
 | 
| -
 | 
| -          // Resulting block must inherit block id from the eliminated current
 | 
| -          // block to guarantee that ordering of phi operands in its successor
 | 
| -          // stays consistent.
 | 
| -          pred->set_block_id(block->block_id());
 | 
| -
 | 
| -          // If v1 and v2 were defined inside eliminated blocks pred1/pred2
 | 
| -          // move them out to the place before inserted IfThenElse instruction.
 | 
| -          EliminateTrivialBlock(pred1, v1->definition(), if_then_else);
 | 
| -          EliminateTrivialBlock(pred2, v2->definition(), if_then_else);
 | 
| -
 | 
| -          // Update use lists to reflect changes in the graph.
 | 
| -          phi->UnuseAllInputs();
 | 
| -          branch->UnuseAllInputs();
 | 
| -          block->UnuseAllInputs();
 | 
| -
 | 
| -          // The graph has changed. Recompute dominators and block orders after
 | 
| -          // this pass is finished.
 | 
| -          changed = true;
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (changed) {
 | 
| -    // We may have changed the block order and the dominator tree.
 | 
| -    flow_graph->DiscoverBlocks();
 | 
| -    GrowableArray<BitVector*> dominance_frontier;
 | 
| -    flow_graph->ComputeDominators(&dominance_frontier);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void FlowGraphOptimizer::EliminateEnvironments() {
 | 
| -  // After this pass we can no longer perform LICM and hoist instructions
 | 
| -  // that can deoptimize.
 | 
| -
 | 
| -  flow_graph_->disallow_licm();
 | 
| -  for (intptr_t i = 0; i < block_order_.length(); ++i) {
 | 
| -    BlockEntryInstr* block = block_order_[i];
 | 
| -    block->RemoveEnvironment();
 | 
| -    for (ForwardInstructionIterator it(block); !it.Done(); it.Advance()) {
 | 
| -      Instruction* current = it.Current();
 | 
| -      if (!current->CanDeoptimize()) {
 | 
| -        // TODO(srdjan): --source-lines needs deopt environments to get at
 | 
| -        // the code for this instruction, however, leaving the environment
 | 
| -        // changes code.
 | 
| -        current->RemoveEnvironment();
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -enum SafeUseCheck { kOptimisticCheck, kStrictCheck };
 | 
| -
 | 
| -// Check if the use is safe for allocation sinking. Allocation sinking
 | 
| -// candidates can only be used at store instructions:
 | 
| -//
 | 
| -//     - any store into the allocation candidate itself is unconditionally safe
 | 
| -//       as it just changes the rematerialization state of this candidate;
 | 
| -//     - store into another object is only safe if another object is allocation
 | 
| -//       candidate.
 | 
| -//
 | 
| -// We use a simple fix-point algorithm to discover the set of valid candidates
 | 
| -// (see CollectCandidates method), that's why this IsSafeUse can operate in two
 | 
| -// modes:
 | 
| -//
 | 
| -//     - optimistic, when every allocation is assumed to be an allocation
 | 
| -//       sinking candidate;
 | 
| -//     - strict, when only marked allocations are assumed to be allocation
 | 
| -//       sinking candidates.
 | 
| -//
 | 
| -// Fix-point algorithm in CollectCandiates first collects a set of allocations
 | 
| -// optimistically and then checks each collected candidate strictly and unmarks
 | 
| -// invalid candidates transitively until only strictly valid ones remain.
 | 
| -static bool IsSafeUse(Value* use, SafeUseCheck check_type) {
 | 
| -  if (use->instruction()->IsMaterializeObject()) {
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  StoreInstanceFieldInstr* store = use->instruction()->AsStoreInstanceField();
 | 
| -  if (store != NULL) {
 | 
| -    if (use == store->value()) {
 | 
| -      Definition* instance = store->instance()->definition();
 | 
| -      return instance->IsAllocateObject() &&
 | 
| -          ((check_type == kOptimisticCheck) ||
 | 
| -           instance->Identity().IsAllocationSinkingCandidate());
 | 
| -    }
 | 
| -    return true;
 | 
| -  }
 | 
| -
 | 
| -  return false;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Right now we are attempting to sink allocation only into
 | 
| -// deoptimization exit. So candidate should only be used in StoreInstanceField
 | 
| -// instructions that write into fields of the allocated object.
 | 
| -// We do not support materialization of the object that has type arguments.
 | 
| -static bool IsAllocationSinkingCandidate(Definition* alloc,
 | 
| -                                         SafeUseCheck check_type) {
 | 
| -  for (Value* use = alloc->input_use_list();
 | 
| -       use != NULL;
 | 
| -       use = use->next_use()) {
 | 
| -    if (!IsSafeUse(use, check_type)) {
 | 
| -      if (FLAG_trace_optimization) {
 | 
| -        OS::Print("use of %s at %s is unsafe for allocation sinking\n",
 | 
| -                  alloc->ToCString(),
 | 
| -                  use->instruction()->ToCString());
 | 
| -      }
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// If the given use is a store into an object then return an object we are
 | 
| -// storing into.
 | 
| -static Definition* StoreInto(Value* use) {
 | 
| -  StoreInstanceFieldInstr* store = use->instruction()->AsStoreInstanceField();
 | 
| -  if (store != NULL) {
 | 
| -    return store->instance()->definition();
 | 
| -  }
 | 
| -
 | 
| -  return NULL;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Remove the given allocation from the graph. It is not observable.
 | 
| -// If deoptimization occurs the object will be materialized.
 | 
| -void AllocationSinking::EliminateAllocation(Definition* alloc) {
 | 
| -  ASSERT(IsAllocationSinkingCandidate(alloc, kStrictCheck));
 | 
| -
 | 
| -  if (FLAG_trace_optimization) {
 | 
| -    OS::Print("removing allocation from the graph: v%" Pd "\n",
 | 
| -              alloc->ssa_temp_index());
 | 
| -  }
 | 
| -
 | 
| -  // As an allocation sinking candidate it is only used in stores to its own
 | 
| -  // fields. Remove these stores.
 | 
| -  for (Value* use = alloc->input_use_list();
 | 
| -       use != NULL;
 | 
| -       use = alloc->input_use_list()) {
 | 
| -    use->instruction()->RemoveFromGraph();
 | 
| -  }
 | 
| -
 | 
| -  // There should be no environment uses. The pass replaced them with
 | 
| -  // MaterializeObject instructions.
 | 
| -#ifdef DEBUG
 | 
| -  for (Value* use = alloc->env_use_list();
 | 
| -       use != NULL;
 | 
| -       use = use->next_use()) {
 | 
| -    ASSERT(use->instruction()->IsMaterializeObject());
 | 
| -  }
 | 
| -#endif
 | 
| -  ASSERT(alloc->input_use_list() == NULL);
 | 
| -  alloc->RemoveFromGraph();
 | 
| -  if (alloc->ArgumentCount() > 0) {
 | 
| -    ASSERT(alloc->ArgumentCount() == 1);
 | 
| -    for (intptr_t i = 0; i < alloc->ArgumentCount(); ++i) {
 | 
| -      alloc->PushArgumentAt(i)->RemoveFromGraph();
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Find allocation instructions that can be potentially eliminated and
 | 
| -// rematerialized at deoptimization exits if needed. See IsSafeUse
 | 
| -// for the description of algorithm used below.
 | 
| -void AllocationSinking::CollectCandidates() {
 | 
| -  // Optimistically collect all potential candidates.
 | 
| -  for (BlockIterator block_it = flow_graph_->reverse_postorder_iterator();
 | 
| -       !block_it.Done();
 | 
| -       block_it.Advance()) {
 | 
| -    BlockEntryInstr* block = block_it.Current();
 | 
| -    for (ForwardInstructionIterator it(block); !it.Done(); it.Advance()) {
 | 
| -      { AllocateObjectInstr* alloc = it.Current()->AsAllocateObject();
 | 
| -        if ((alloc != NULL) &&
 | 
| -            IsAllocationSinkingCandidate(alloc, kOptimisticCheck)) {
 | 
| -          alloc->SetIdentity(AliasIdentity::AllocationSinkingCandidate());
 | 
| -          candidates_.Add(alloc);
 | 
| -        }
 | 
| -      }
 | 
| -      { AllocateUninitializedContextInstr* alloc =
 | 
| -            it.Current()->AsAllocateUninitializedContext();
 | 
| -        if ((alloc != NULL) &&
 | 
| -            IsAllocationSinkingCandidate(alloc, kOptimisticCheck)) {
 | 
| -          alloc->SetIdentity(AliasIdentity::AllocationSinkingCandidate());
 | 
| -          candidates_.Add(alloc);
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Transitively unmark all candidates that are not strictly valid.
 | 
| -  bool changed;
 | 
| -  do {
 | 
| -    changed = false;
 | 
| -    for (intptr_t i = 0; i < candidates_.length(); i++) {
 | 
| -      Definition* alloc = candidates_[i];
 | 
| -      if (alloc->Identity().IsAllocationSinkingCandidate()) {
 | 
| -        if (!IsAllocationSinkingCandidate(alloc, kStrictCheck)) {
 | 
| -          alloc->SetIdentity(AliasIdentity::Unknown());
 | 
| -          changed = true;
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  } while (changed);
 | 
| -
 | 
| -  // Shrink the list of candidates removing all unmarked ones.
 | 
| -  intptr_t j = 0;
 | 
| -  for (intptr_t i = 0; i < candidates_.length(); i++) {
 | 
| -    Definition* alloc = candidates_[i];
 | 
| -    if (alloc->Identity().IsAllocationSinkingCandidate()) {
 | 
| -      if (FLAG_trace_optimization) {
 | 
| -        OS::Print("discovered allocation sinking candidate: v%" Pd "\n",
 | 
| -                  alloc->ssa_temp_index());
 | 
| -      }
 | 
| -
 | 
| -      if (j != i) {
 | 
| -        candidates_[j] = alloc;
 | 
| -      }
 | 
| -      j++;
 | 
| -    }
 | 
| -  }
 | 
| -  candidates_.TruncateTo(j);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// If materialization references an allocation sinking candidate then replace
 | 
| -// this reference with a materialization which should have been computed for
 | 
| -// this side-exit. CollectAllExits should have collected this exit.
 | 
| -void AllocationSinking::NormalizeMaterializations() {
 | 
| -  for (intptr_t i = 0; i < candidates_.length(); i++) {
 | 
| -    Definition* alloc = candidates_[i];
 | 
| -
 | 
| -    Value* next_use;
 | 
| -    for (Value* use = alloc->input_use_list();
 | 
| -         use != NULL;
 | 
| -         use = next_use) {
 | 
| -      next_use = use->next_use();
 | 
| -      if (use->instruction()->IsMaterializeObject()) {
 | 
| -        use->BindTo(MaterializationFor(alloc, use->instruction()));
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// We transitively insert materializations at each deoptimization exit that
 | 
| -// might see the given allocation (see ExitsCollector). Some of this
 | 
| -// materializations are not actually used and some fail to compute because
 | 
| -// they are inserted in the block that is not dominated by the allocation.
 | 
| -// Remove them unused materializations from the graph.
 | 
| -void AllocationSinking::RemoveUnusedMaterializations() {
 | 
| -  intptr_t j = 0;
 | 
| -  for (intptr_t i = 0; i < materializations_.length(); i++) {
 | 
| -    MaterializeObjectInstr* mat = materializations_[i];
 | 
| -    if ((mat->input_use_list() == NULL) && (mat->env_use_list() == NULL)) {
 | 
| -      // Check if this materialization failed to compute and remove any
 | 
| -      // unforwarded loads. There were no loads from any allocation sinking
 | 
| -      // candidate in the beggining so it is safe to assume that any encountered
 | 
| -      // load was inserted by CreateMaterializationAt.
 | 
| -      for (intptr_t i = 0; i < mat->InputCount(); i++) {
 | 
| -        LoadFieldInstr* load = mat->InputAt(i)->definition()->AsLoadField();
 | 
| -        if ((load != NULL) &&
 | 
| -            (load->instance()->definition() == mat->allocation())) {
 | 
| -          load->ReplaceUsesWith(flow_graph_->constant_null());
 | 
| -          load->RemoveFromGraph();
 | 
| -        }
 | 
| -      }
 | 
| -      mat->RemoveFromGraph();
 | 
| -    } else {
 | 
| -      if (j != i) {
 | 
| -        materializations_[j] = mat;
 | 
| -      }
 | 
| -      j++;
 | 
| -    }
 | 
| -  }
 | 
| -  materializations_.TruncateTo(j);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Some candidates might stop being eligible for allocation sinking after
 | 
| -// the load forwarding because they flow into phis that load forwarding
 | 
| -// inserts. Discover such allocations and remove them from the list
 | 
| -// of allocation sinking candidates undoing all changes that we did
 | 
| -// in preparation for sinking these allocations.
 | 
| -void AllocationSinking::DiscoverFailedCandidates() {
 | 
| -  // Transitively unmark all candidates that are not strictly valid.
 | 
| -  bool changed;
 | 
| -  do {
 | 
| -    changed = false;
 | 
| -    for (intptr_t i = 0; i < candidates_.length(); i++) {
 | 
| -      Definition* alloc = candidates_[i];
 | 
| -      if (alloc->Identity().IsAllocationSinkingCandidate()) {
 | 
| -        if (!IsAllocationSinkingCandidate(alloc, kStrictCheck)) {
 | 
| -          alloc->SetIdentity(AliasIdentity::Unknown());
 | 
| -          changed = true;
 | 
| -        }
 | 
| -      }
 | 
| -    }
 | 
| -  } while (changed);
 | 
| -
 | 
| -  // Remove all failed candidates from the candidates list.
 | 
| -  intptr_t j = 0;
 | 
| -  for (intptr_t i = 0; i < candidates_.length(); i++) {
 | 
| -    Definition* alloc = candidates_[i];
 | 
| -    if (!alloc->Identity().IsAllocationSinkingCandidate()) {
 | 
| -      if (FLAG_trace_optimization) {
 | 
| -        OS::Print("allocation v%" Pd " can't be eliminated\n",
 | 
| -                  alloc->ssa_temp_index());
 | 
| -      }
 | 
| -
 | 
| -#ifdef DEBUG
 | 
| -      for (Value* use = alloc->env_use_list();
 | 
| -           use != NULL;
 | 
| -           use = use->next_use()) {
 | 
| -        ASSERT(use->instruction()->IsMaterializeObject());
 | 
| -      }
 | 
| -#endif
 | 
| -
 | 
| -      // All materializations will be removed from the graph. Remove inserted
 | 
| -      // loads first and detach materializations from allocation's environment
 | 
| -      // use list: we will reconstruct it when we start removing
 | 
| -      // materializations.
 | 
| -      alloc->set_env_use_list(NULL);
 | 
| -      for (Value* use = alloc->input_use_list();
 | 
| -           use != NULL;
 | 
| -           use = use->next_use()) {
 | 
| -        if (use->instruction()->IsLoadField()) {
 | 
| -          LoadFieldInstr* load = use->instruction()->AsLoadField();
 | 
| -          load->ReplaceUsesWith(flow_graph_->constant_null());
 | 
| -          load->RemoveFromGraph();
 | 
| -        } else {
 | 
| -          ASSERT(use->instruction()->IsMaterializeObject() ||
 | 
| -                 use->instruction()->IsPhi() ||
 | 
| -                 use->instruction()->IsStoreInstanceField());
 | 
| -        }
 | 
| -      }
 | 
| -    } else {
 | 
| -      if (j != i) {
 | 
| -        candidates_[j] = alloc;
 | 
| -      }
 | 
| -      j++;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (j != candidates_.length()) {  // Something was removed from candidates.
 | 
| -    intptr_t k = 0;
 | 
| -    for (intptr_t i = 0; i < materializations_.length(); i++) {
 | 
| -      MaterializeObjectInstr* mat = materializations_[i];
 | 
| -      if (!mat->allocation()->Identity().IsAllocationSinkingCandidate()) {
 | 
| -        // Restore environment uses of the allocation that were replaced
 | 
| -        // by this materialization and drop materialization.
 | 
| -        mat->ReplaceUsesWith(mat->allocation());
 | 
| -        mat->RemoveFromGraph();
 | 
| -      } else {
 | 
| -        if (k != i) {
 | 
| -          materializations_[k] = mat;
 | 
| -        }
 | 
| -        k++;
 | 
| -      }
 | 
| -    }
 | 
| -    materializations_.TruncateTo(k);
 | 
| -  }
 | 
| -
 | 
| -  candidates_.TruncateTo(j);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void AllocationSinking::Optimize() {
 | 
| -  CollectCandidates();
 | 
| -
 | 
| -  // Insert MaterializeObject instructions that will describe the state of the
 | 
| -  // object at all deoptimization points. Each inserted materialization looks
 | 
| -  // like this (where v_0 is allocation that we are going to eliminate):
 | 
| -  //   v_1     <- LoadField(v_0, field_1)
 | 
| -  //           ...
 | 
| -  //   v_N     <- LoadField(v_0, field_N)
 | 
| -  //   v_{N+1} <- MaterializeObject(field_1 = v_1, ..., field_N = v_{N})
 | 
| -  for (intptr_t i = 0; i < candidates_.length(); i++) {
 | 
| -    InsertMaterializations(candidates_[i]);
 | 
| -  }
 | 
| -
 | 
| -  // Run load forwarding to eliminate LoadField instructions inserted above.
 | 
| -  // All loads will be successfully eliminated because:
 | 
| -  //   a) they use fields (not offsets) and thus provide precise aliasing
 | 
| -  //      information
 | 
| -  //   b) candidate does not escape and thus its fields is not affected by
 | 
| -  //      external effects from calls.
 | 
| -  LoadOptimizer::OptimizeGraph(flow_graph_);
 | 
| -
 | 
| -  NormalizeMaterializations();
 | 
| -
 | 
| -  RemoveUnusedMaterializations();
 | 
| -
 | 
| -  // If any candidates are no longer eligible for allocation sinking abort
 | 
| -  // the optimization for them and undo any changes we did in preparation.
 | 
| -  DiscoverFailedCandidates();
 | 
| -
 | 
| -  // At this point we have computed the state of object at each deoptimization
 | 
| -  // point and we can eliminate it. Loads inserted above were forwarded so there
 | 
| -  // are no uses of the allocation just as in the begging of the pass.
 | 
| -  for (intptr_t i = 0; i < candidates_.length(); i++) {
 | 
| -    EliminateAllocation(candidates_[i]);
 | 
| -  }
 | 
| -
 | 
| -  // Process materializations and unbox their arguments: materializations
 | 
| -  // are part of the environment and can materialize boxes for double/mint/simd
 | 
| -  // values when needed.
 | 
| -  // TODO(vegorov): handle all box types here.
 | 
| -  for (intptr_t i = 0; i < materializations_.length(); i++) {
 | 
| -    MaterializeObjectInstr* mat = materializations_[i];
 | 
| -    for (intptr_t j = 0; j < mat->InputCount(); j++) {
 | 
| -      Definition* defn = mat->InputAt(j)->definition();
 | 
| -      if (defn->IsBox()) {
 | 
| -        mat->InputAt(j)->BindTo(defn->InputAt(0)->definition());
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Remove materializations from the graph. Register allocator will treat them
 | 
| -// as part of the environment not as a real instruction.
 | 
| -void AllocationSinking::DetachMaterializations() {
 | 
| -  for (intptr_t i = 0; i < materializations_.length(); i++) {
 | 
| -    materializations_[i]->previous()->LinkTo(materializations_[i]->next());
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Add a field/offset to the list of fields if it is not yet present there.
 | 
| -static bool AddSlot(ZoneGrowableArray<const Object*>* slots,
 | 
| -                    const Object& slot) {
 | 
| -  for (intptr_t i = 0; i < slots->length(); i++) {
 | 
| -    if ((*slots)[i]->raw() == slot.raw()) {
 | 
| -      return false;
 | 
| -    }
 | 
| -  }
 | 
| -  slots->Add(&slot);
 | 
| -  return true;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Find deoptimization exit for the given materialization assuming that all
 | 
| -// materializations are emitted right before the instruction which is a
 | 
| -// deoptimization exit.
 | 
| -static Instruction* ExitForMaterialization(MaterializeObjectInstr* mat) {
 | 
| -  while (mat->next()->IsMaterializeObject()) {
 | 
| -    mat = mat->next()->AsMaterializeObject();
 | 
| -  }
 | 
| -  return mat->next();
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Given the deoptimization exit find first materialization that was inserted
 | 
| -// before it.
 | 
| -static Instruction* FirstMaterializationAt(Instruction* exit) {
 | 
| -  while (exit->previous()->IsMaterializeObject()) {
 | 
| -    exit = exit->previous();
 | 
| -  }
 | 
| -  return exit;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Given the allocation and deoptimization exit try to find MaterializeObject
 | 
| -// instruction corresponding to this allocation at this exit.
 | 
| -MaterializeObjectInstr* AllocationSinking::MaterializationFor(
 | 
| -    Definition* alloc, Instruction* exit) {
 | 
| -  if (exit->IsMaterializeObject()) {
 | 
| -    exit = ExitForMaterialization(exit->AsMaterializeObject());
 | 
| -  }
 | 
| -
 | 
| -  for (MaterializeObjectInstr* mat = exit->previous()->AsMaterializeObject();
 | 
| -       mat != NULL;
 | 
| -       mat = mat->previous()->AsMaterializeObject()) {
 | 
| -    if (mat->allocation() == alloc) {
 | 
| -      return mat;
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  return NULL;
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Insert MaterializeObject instruction for the given allocation before
 | 
| -// the given instruction that can deoptimize.
 | 
| -void AllocationSinking::CreateMaterializationAt(
 | 
| -    Instruction* exit,
 | 
| -    Definition* alloc,
 | 
| -    const ZoneGrowableArray<const Object*>& slots) {
 | 
| -  ZoneGrowableArray<Value*>* values =
 | 
| -      new(I) ZoneGrowableArray<Value*>(slots.length());
 | 
| -
 | 
| -  // All loads should be inserted before the first materialization so that
 | 
| -  // IR follows the following pattern: loads, materializations, deoptimizing
 | 
| -  // instruction.
 | 
| -  Instruction* load_point = FirstMaterializationAt(exit);
 | 
| -
 | 
| -  // Insert load instruction for every field.
 | 
| -  for (intptr_t i = 0; i < slots.length(); i++) {
 | 
| -    LoadFieldInstr* load = slots[i]->IsField()
 | 
| -        ? new(I) LoadFieldInstr(
 | 
| -            new(I) Value(alloc),
 | 
| -            &Field::Cast(*slots[i]),
 | 
| -            AbstractType::ZoneHandle(I),
 | 
| -            alloc->token_pos())
 | 
| -        : new(I) LoadFieldInstr(
 | 
| -            new(I) Value(alloc),
 | 
| -            Smi::Cast(*slots[i]).Value(),
 | 
| -            AbstractType::ZoneHandle(I),
 | 
| -            alloc->token_pos());
 | 
| -    flow_graph_->InsertBefore(
 | 
| -        load_point, load, NULL, FlowGraph::kValue);
 | 
| -    values->Add(new(I) Value(load));
 | 
| -  }
 | 
| -
 | 
| -  MaterializeObjectInstr* mat = NULL;
 | 
| -  if (alloc->IsAllocateObject()) {
 | 
| -    mat = new(I) MaterializeObjectInstr(
 | 
| -        alloc->AsAllocateObject(), slots, values);
 | 
| -  } else {
 | 
| -    ASSERT(alloc->IsAllocateUninitializedContext());
 | 
| -    mat = new(I) MaterializeObjectInstr(
 | 
| -        alloc->AsAllocateUninitializedContext(), slots, values);
 | 
| -  }
 | 
| -
 | 
| -  flow_graph_->InsertBefore(exit, mat, NULL, FlowGraph::kValue);
 | 
| -
 | 
| -  // Replace all mentions of this allocation with a newly inserted
 | 
| -  // MaterializeObject instruction.
 | 
| -  // We must preserve the identity: all mentions are replaced by the same
 | 
| -  // materialization.
 | 
| -  for (Environment::DeepIterator env_it(exit->env());
 | 
| -       !env_it.Done();
 | 
| -       env_it.Advance()) {
 | 
| -    Value* use = env_it.CurrentValue();
 | 
| -    if (use->definition() == alloc) {
 | 
| -      use->RemoveFromUseList();
 | 
| -      use->set_definition(mat);
 | 
| -      mat->AddEnvUse(use);
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Mark MaterializeObject as an environment use of this allocation.
 | 
| -  // This will allow us to discover it when we are looking for deoptimization
 | 
| -  // exits for another allocation that potentially flows into this one.
 | 
| -  Value* val = new(I) Value(alloc);
 | 
| -  val->set_instruction(mat);
 | 
| -  alloc->AddEnvUse(val);
 | 
| -
 | 
| -  // Record inserted materialization.
 | 
| -  materializations_.Add(mat);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Add given instruction to the list of the instructions if it is not yet
 | 
| -// present there.
 | 
| -template<typename T>
 | 
| -void AddInstruction(GrowableArray<T*>* list, T* value) {
 | 
| -  ASSERT(!value->IsGraphEntry());
 | 
| -  for (intptr_t i = 0; i < list->length(); i++) {
 | 
| -    if ((*list)[i] == value) {
 | 
| -      return;
 | 
| -    }
 | 
| -  }
 | 
| -  list->Add(value);
 | 
| -}
 | 
| -
 | 
| -
 | 
| -// Transitively collect all deoptimization exits that might need this allocation
 | 
| -// rematerialized. It is not enough to collect only environment uses of this
 | 
| -// allocation because it can flow into other objects that will be
 | 
| -// dematerialized and that are referenced by deopt environments that
 | 
| -// don't contain this allocation explicitly.
 | 
| -void AllocationSinking::ExitsCollector::Collect(Definition* alloc) {
 | 
| -  for (Value* use = alloc->env_use_list();
 | 
| -       use != NULL;
 | 
| -       use = use->next_use()) {
 | 
| -    if (use->instruction()->IsMaterializeObject()) {
 | 
| -      AddInstruction(&exits_, ExitForMaterialization(
 | 
| -          use->instruction()->AsMaterializeObject()));
 | 
| -    } else {
 | 
| -      AddInstruction(&exits_, use->instruction());
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  // Check if this allocation is stored into any other allocation sinking
 | 
| -  // candidate and put it on worklist so that we conservatively collect all
 | 
| -  // exits for that candidate as well because they potentially might see
 | 
| -  // this object.
 | 
| -  for (Value* use = alloc->input_use_list();
 | 
| -       use != NULL;
 | 
| -       use = use->next_use()) {
 | 
| -    Definition* obj = StoreInto(use);
 | 
| -    if ((obj != NULL) && (obj != alloc)) {
 | 
| -      AddInstruction(&worklist_, obj);
 | 
| -    }
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void AllocationSinking::ExitsCollector::CollectTransitively(Definition* alloc) {
 | 
| -  exits_.TruncateTo(0);
 | 
| -  worklist_.TruncateTo(0);
 | 
| -
 | 
| -  worklist_.Add(alloc);
 | 
| -
 | 
| -  // Note: worklist potentially will grow while we are iterating over it.
 | 
| -  // We are not removing allocations from the worklist not to waste space on
 | 
| -  // the side maintaining BitVector of already processed allocations: worklist
 | 
| -  // is expected to be very small thus linear search in it is just as effecient
 | 
| -  // as a bitvector.
 | 
| -  for (intptr_t i = 0; i < worklist_.length(); i++) {
 | 
| -    Collect(worklist_[i]);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
| -void AllocationSinking::InsertMaterializations(Definition* alloc) {
 | 
| -  // Collect all fields that are written for this instance.
 | 
| -  ZoneGrowableArray<const Object*>* slots =
 | 
| -      new(I) ZoneGrowableArray<const Object*>(5);
 | 
| -
 | 
| -  for (Value* use = alloc->input_use_list();
 | 
| -       use != NULL;
 | 
| -       use = use->next_use()) {
 | 
| -    StoreInstanceFieldInstr* store = use->instruction()->AsStoreInstanceField();
 | 
| -    if ((store != NULL) && (store->instance()->definition() == alloc)) {
 | 
| -      if (!store->field().IsNull()) {
 | 
| -        AddSlot(slots, store->field());
 | 
| -      } else {
 | 
| -        AddSlot(slots, Smi::ZoneHandle(I, Smi::New(store->offset_in_bytes())));
 | 
| -      }
 | 
| -    }
 | 
| -  }
 | 
| -
 | 
| -  if (alloc->ArgumentCount() > 0) {
 | 
| -    AllocateObjectInstr* alloc_object = alloc->AsAllocateObject();
 | 
| -    ASSERT(alloc_object->ArgumentCount() == 1);
 | 
| -    intptr_t type_args_offset =
 | 
| -        alloc_object->cls().type_arguments_field_offset();
 | 
| -    AddSlot(slots, Smi::ZoneHandle(I, Smi::New(type_args_offset)));
 | 
| -  }
 | 
| -
 | 
| -  // Collect all instructions that mention this object in the environment.
 | 
| -  exits_collector_.CollectTransitively(alloc);
 | 
| -
 | 
| -  // Insert materializations at environment uses.
 | 
| -  for (intptr_t i = 0; i < exits_collector_.exits().length(); i++) {
 | 
| -    CreateMaterializationAt(
 | 
| -        exits_collector_.exits()[i], alloc, *slots);
 | 
| -  }
 | 
| -}
 | 
| -
 | 
| -
 | 
|  }  // namespace dart
 | 
| 
 |