Index: runtime/vm/constant_propagator.cc |
=================================================================== |
--- runtime/vm/constant_propagator.cc (revision 42566) |
+++ runtime/vm/constant_propagator.cc (working copy) |
@@ -2,7504 +2,27 @@ |
// for details. All rights reserved. Use of this source code is governed by a |
// BSD-style license that can be found in the LICENSE file. |
-#include "vm/flow_graph_optimizer.h" |
+#include "vm/constant_propagator.h" |
#include "vm/bit_vector.h" |
-#include "vm/cha.h" |
-#include "vm/cpu.h" |
-#include "vm/dart_entry.h" |
-#include "vm/exceptions.h" |
#include "vm/flow_graph_builder.h" |
#include "vm/flow_graph_compiler.h" |
#include "vm/flow_graph_range_analysis.h" |
-#include "vm/hash_map.h" |
#include "vm/il_printer.h" |
#include "vm/intermediate_language.h" |
-#include "vm/object_store.h" |
#include "vm/parser.h" |
-#include "vm/resolver.h" |
-#include "vm/scopes.h" |
-#include "vm/stack_frame.h" |
#include "vm/symbols.h" |
namespace dart { |
-DEFINE_FLAG(int, getter_setter_ratio, 13, |
- "Ratio of getter/setter usage used for double field unboxing heuristics"); |
-DEFINE_FLAG(bool, load_cse, true, "Use redundant load elimination."); |
-DEFINE_FLAG(bool, dead_store_elimination, true, "Eliminate dead stores"); |
-DEFINE_FLAG(int, max_polymorphic_checks, 4, |
- "Maximum number of polymorphic check, otherwise it is megamorphic."); |
-DEFINE_FLAG(int, max_equality_polymorphic_checks, 32, |
- "Maximum number of polymorphic checks in equality operator," |
- " otherwise use megamorphic dispatch."); |
DEFINE_FLAG(bool, remove_redundant_phis, true, "Remove redundant phis."); |
DEFINE_FLAG(bool, trace_constant_propagation, false, |
"Print constant propagation and useless code elimination."); |
-DEFINE_FLAG(bool, trace_load_optimization, false, |
- "Print live sets for load optimization pass."); |
-DEFINE_FLAG(bool, trace_optimization, false, "Print optimization details."); |
-DEFINE_FLAG(bool, truncating_left_shift, true, |
- "Optimize left shift to truncate if possible"); |
-DEFINE_FLAG(bool, use_cha, true, "Use class hierarchy analysis."); |
-#if defined(TARGET_ARCH_ARM) || defined(TARGET_ARCH_IA32) |
-DEFINE_FLAG(bool, trace_smi_widening, false, "Trace Smi->Int32 widening pass."); |
-#endif |
-DECLARE_FLAG(bool, enable_type_checks); |
-DECLARE_FLAG(bool, source_lines); |
-DECLARE_FLAG(bool, trace_type_check_elimination); |
-DECLARE_FLAG(bool, warn_on_javascript_compatibility); |
// Quick access to the locally defined isolate() method. |
#define I (isolate()) |
-static bool ShouldInlineSimd() { |
- return FlowGraphCompiler::SupportsUnboxedSimd128(); |
-} |
- |
-static bool CanUnboxDouble() { |
- return FlowGraphCompiler::SupportsUnboxedDoubles(); |
-} |
- |
- |
-static bool ShouldInlineInt64ArrayOps() { |
-#if defined(TARGET_ARCH_X64) |
- return true; |
-#endif |
- return false; |
-} |
- |
-static bool CanConvertUnboxedMintToDouble() { |
-#if defined(TARGET_ARCH_IA32) |
- return true; |
-#else |
- // ARM does not have a short instruction sequence for converting int64 to |
- // double. |
- // TODO(johnmccutchan): Investigate possibility on MIPS once |
- // mints are implemented there. |
- return false; |
-#endif |
-} |
- |
- |
-// Optimize instance calls using ICData. |
-void FlowGraphOptimizer::ApplyICData() { |
- VisitBlocks(); |
-} |
- |
- |
-// Optimize instance calls using cid. This is called after optimizer |
-// converted instance calls to instructions. Any remaining |
-// instance calls are either megamorphic calls, cannot be optimized or |
-// have no runtime type feedback collected. |
-// Attempts to convert an instance call (IC call) using propagated class-ids, |
-// e.g., receiver class id, guarded-cid, or by guessing cid-s. |
-void FlowGraphOptimizer::ApplyClassIds() { |
- ASSERT(current_iterator_ == NULL); |
- for (intptr_t i = 0; i < block_order_.length(); ++i) { |
- BlockEntryInstr* entry = block_order_[i]; |
- ForwardInstructionIterator it(entry); |
- current_iterator_ = ⁢ |
- for (; !it.Done(); it.Advance()) { |
- Instruction* instr = it.Current(); |
- if (instr->IsInstanceCall()) { |
- InstanceCallInstr* call = instr->AsInstanceCall(); |
- if (call->HasICData()) { |
- if (TryCreateICData(call)) { |
- VisitInstanceCall(call); |
- } |
- } |
- } else if (instr->IsPolymorphicInstanceCall()) { |
- SpecializePolymorphicInstanceCall(instr->AsPolymorphicInstanceCall()); |
- } else if (instr->IsStrictCompare()) { |
- VisitStrictCompare(instr->AsStrictCompare()); |
- } else if (instr->IsBranch()) { |
- ComparisonInstr* compare = instr->AsBranch()->comparison(); |
- if (compare->IsStrictCompare()) { |
- VisitStrictCompare(compare->AsStrictCompare()); |
- } |
- } |
- } |
- current_iterator_ = NULL; |
- } |
-} |
- |
- |
-// TODO(srdjan): Test/support other number types as well. |
-static bool IsNumberCid(intptr_t cid) { |
- return (cid == kSmiCid) || (cid == kDoubleCid); |
-} |
- |
- |
-// Attempt to build ICData for call using propagated class-ids. |
-bool FlowGraphOptimizer::TryCreateICData(InstanceCallInstr* call) { |
- ASSERT(call->HasICData()); |
- if (call->ic_data()->NumberOfUsedChecks() > 0) { |
- // This occurs when an instance call has too many checks, will be converted |
- // to megamorphic call. |
- return false; |
- } |
- if (FLAG_warn_on_javascript_compatibility) { |
- // Do not make the instance call megamorphic if the callee needs to decode |
- // the calling code sequence to lookup the ic data and verify if a warning |
- // has already been issued or not. |
- // TryCreateICData is only invoked if the ic_data target has not been called |
- // yet, so no warning can possibly have been issued. |
- ASSERT(!call->ic_data()->IssuedJSWarning()); |
- if (call->ic_data()->MayCheckForJSWarning()) { |
- return false; |
- } |
- } |
- GrowableArray<intptr_t> class_ids(call->ic_data()->NumArgsTested()); |
- ASSERT(call->ic_data()->NumArgsTested() <= call->ArgumentCount()); |
- for (intptr_t i = 0; i < call->ic_data()->NumArgsTested(); i++) { |
- const intptr_t cid = call->PushArgumentAt(i)->value()->Type()->ToCid(); |
- class_ids.Add(cid); |
- } |
- |
- const Token::Kind op_kind = call->token_kind(); |
- if (Token::IsRelationalOperator(op_kind) || |
- Token::IsEqualityOperator(op_kind) || |
- Token::IsBinaryOperator(op_kind)) { |
- // Guess cid: if one of the inputs is a number assume that the other |
- // is a number of same type. |
- const intptr_t cid_0 = class_ids[0]; |
- const intptr_t cid_1 = class_ids[1]; |
- if ((cid_0 == kDynamicCid) && (IsNumberCid(cid_1))) { |
- class_ids[0] = cid_1; |
- } else if (IsNumberCid(cid_0) && (cid_1 == kDynamicCid)) { |
- class_ids[1] = cid_0; |
- } |
- } |
- |
- for (intptr_t i = 0; i < class_ids.length(); i++) { |
- if (class_ids[i] == kDynamicCid) { |
- // Not all cid-s known. |
- return false; |
- } |
- } |
- |
- const Array& args_desc_array = Array::Handle(I, |
- ArgumentsDescriptor::New(call->ArgumentCount(), call->argument_names())); |
- ArgumentsDescriptor args_desc(args_desc_array); |
- const Class& receiver_class = Class::Handle(I, |
- isolate()->class_table()->At(class_ids[0])); |
- const Function& function = Function::Handle(I, |
- Resolver::ResolveDynamicForReceiverClass( |
- receiver_class, |
- call->function_name(), |
- args_desc)); |
- if (function.IsNull()) { |
- return false; |
- } |
- // Create new ICData, do not modify the one attached to the instruction |
- // since it is attached to the assembly instruction itself. |
- // TODO(srdjan): Prevent modification of ICData object that is |
- // referenced in assembly code. |
- ICData& ic_data = ICData::ZoneHandle(I, ICData::New( |
- flow_graph_->parsed_function()->function(), |
- call->function_name(), |
- args_desc_array, |
- call->deopt_id(), |
- class_ids.length())); |
- if (class_ids.length() > 1) { |
- ic_data.AddCheck(class_ids, function); |
- } else { |
- ASSERT(class_ids.length() == 1); |
- ic_data.AddReceiverCheck(class_ids[0], function); |
- } |
- call->set_ic_data(&ic_data); |
- return true; |
-} |
- |
- |
-const ICData& FlowGraphOptimizer::TrySpecializeICData(const ICData& ic_data, |
- intptr_t cid) { |
- ASSERT(ic_data.NumArgsTested() == 1); |
- |
- if ((ic_data.NumberOfUsedChecks() == 1) && ic_data.HasReceiverClassId(cid)) { |
- return ic_data; // Nothing to do |
- } |
- |
- const Function& function = |
- Function::Handle(I, ic_data.GetTargetForReceiverClassId(cid)); |
- // TODO(fschneider): Try looking up the function on the class if it is |
- // not found in the ICData. |
- if (!function.IsNull()) { |
- const ICData& new_ic_data = ICData::ZoneHandle(I, ICData::New( |
- Function::Handle(I, ic_data.owner()), |
- String::Handle(I, ic_data.target_name()), |
- Object::empty_array(), // Dummy argument descriptor. |
- ic_data.deopt_id(), |
- ic_data.NumArgsTested())); |
- new_ic_data.SetDeoptReasons(ic_data.DeoptReasons()); |
- new_ic_data.AddReceiverCheck(cid, function); |
- return new_ic_data; |
- } |
- |
- return ic_data; |
-} |
- |
- |
-void FlowGraphOptimizer::SpecializePolymorphicInstanceCall( |
- PolymorphicInstanceCallInstr* call) { |
- if (!call->with_checks()) { |
- return; // Already specialized. |
- } |
- |
- const intptr_t receiver_cid = |
- call->PushArgumentAt(0)->value()->Type()->ToCid(); |
- if (receiver_cid == kDynamicCid) { |
- return; // No information about receiver was infered. |
- } |
- |
- const ICData& ic_data = TrySpecializeICData(call->ic_data(), receiver_cid); |
- if (ic_data.raw() == call->ic_data().raw()) { |
- // No specialization. |
- return; |
- } |
- |
- const bool with_checks = false; |
- PolymorphicInstanceCallInstr* specialized = |
- new(I) PolymorphicInstanceCallInstr(call->instance_call(), |
- ic_data, |
- with_checks); |
- call->ReplaceWith(specialized, current_iterator()); |
-} |
- |
- |
-static BinarySmiOpInstr* AsSmiShiftLeftInstruction(Definition* d) { |
- BinarySmiOpInstr* instr = d->AsBinarySmiOp(); |
- if ((instr != NULL) && (instr->op_kind() == Token::kSHL)) { |
- return instr; |
- } |
- return NULL; |
-} |
- |
- |
-static bool IsPositiveOrZeroSmiConst(Definition* d) { |
- ConstantInstr* const_instr = d->AsConstant(); |
- if ((const_instr != NULL) && (const_instr->value().IsSmi())) { |
- return Smi::Cast(const_instr->value()).Value() >= 0; |
- } |
- return false; |
-} |
- |
- |
-void FlowGraphOptimizer::OptimizeLeftShiftBitAndSmiOp( |
- Definition* bit_and_instr, |
- Definition* left_instr, |
- Definition* right_instr) { |
- ASSERT(bit_and_instr != NULL); |
- ASSERT((left_instr != NULL) && (right_instr != NULL)); |
- |
- // Check for pattern, smi_shift_left must be single-use. |
- bool is_positive_or_zero = IsPositiveOrZeroSmiConst(left_instr); |
- if (!is_positive_or_zero) { |
- is_positive_or_zero = IsPositiveOrZeroSmiConst(right_instr); |
- } |
- if (!is_positive_or_zero) return; |
- |
- BinarySmiOpInstr* smi_shift_left = NULL; |
- if (bit_and_instr->InputAt(0)->IsSingleUse()) { |
- smi_shift_left = AsSmiShiftLeftInstruction(left_instr); |
- } |
- if ((smi_shift_left == NULL) && (bit_and_instr->InputAt(1)->IsSingleUse())) { |
- smi_shift_left = AsSmiShiftLeftInstruction(right_instr); |
- } |
- if (smi_shift_left == NULL) return; |
- |
- // Pattern recognized. |
- smi_shift_left->mark_truncating(); |
- ASSERT(bit_and_instr->IsBinarySmiOp() || bit_and_instr->IsBinaryMintOp()); |
- if (bit_and_instr->IsBinaryMintOp()) { |
- // Replace Mint op with Smi op. |
- BinarySmiOpInstr* smi_op = new(I) BinarySmiOpInstr( |
- Token::kBIT_AND, |
- new(I) Value(left_instr), |
- new(I) Value(right_instr), |
- Isolate::kNoDeoptId); // BIT_AND cannot deoptimize. |
- bit_and_instr->ReplaceWith(smi_op, current_iterator()); |
- } |
-} |
- |
- |
- |
-// Used by TryMergeDivMod. |
-// Inserts a load-indexed instruction between a TRUNCDIV or MOD instruction, |
-// and the using instruction. This is an intermediate step before merging. |
-void FlowGraphOptimizer::AppendLoadIndexedForMerged(Definition* instr, |
- intptr_t ix, |
- intptr_t cid) { |
- const intptr_t index_scale = Instance::ElementSizeFor(cid); |
- ConstantInstr* index_instr = |
- flow_graph()->GetConstant(Smi::Handle(I, Smi::New(ix))); |
- LoadIndexedInstr* load = |
- new(I) LoadIndexedInstr(new(I) Value(instr), |
- new(I) Value(index_instr), |
- index_scale, |
- cid, |
- Isolate::kNoDeoptId, |
- instr->token_pos()); |
- instr->ReplaceUsesWith(load); |
- flow_graph()->InsertAfter(instr, load, NULL, FlowGraph::kValue); |
-} |
- |
- |
-void FlowGraphOptimizer::AppendExtractNthOutputForMerged(Definition* instr, |
- intptr_t index, |
- Representation rep, |
- intptr_t cid) { |
- ExtractNthOutputInstr* extract = |
- new(I) ExtractNthOutputInstr(new(I) Value(instr), index, rep, cid); |
- instr->ReplaceUsesWith(extract); |
- flow_graph()->InsertAfter(instr, extract, NULL, FlowGraph::kValue); |
-} |
- |
- |
-// Dart: |
-// var x = d % 10; |
-// var y = d ~/ 10; |
-// var z = x + y; |
-// |
-// IL: |
-// v4 <- %(v2, v3) |
-// v5 <- ~/(v2, v3) |
-// v6 <- +(v4, v5) |
-// |
-// IL optimized: |
-// v4 <- DIVMOD(v2, v3); |
-// v5 <- LoadIndexed(v4, 0); // ~/ result |
-// v6 <- LoadIndexed(v4, 1); // % result |
-// v7 <- +(v5, v6) |
-// Because of the environment it is important that merged instruction replaces |
-// first original instruction encountered. |
-void FlowGraphOptimizer::TryMergeTruncDivMod( |
- GrowableArray<BinarySmiOpInstr*>* merge_candidates) { |
- if (merge_candidates->length() < 2) { |
- // Need at least a TRUNCDIV and a MOD. |
- return; |
- } |
- for (intptr_t i = 0; i < merge_candidates->length(); i++) { |
- BinarySmiOpInstr* curr_instr = (*merge_candidates)[i]; |
- if (curr_instr == NULL) { |
- // Instruction was merged already. |
- continue; |
- } |
- ASSERT((curr_instr->op_kind() == Token::kTRUNCDIV) || |
- (curr_instr->op_kind() == Token::kMOD)); |
- // Check if there is kMOD/kTRUNDIV binop with same inputs. |
- const intptr_t other_kind = (curr_instr->op_kind() == Token::kTRUNCDIV) ? |
- Token::kMOD : Token::kTRUNCDIV; |
- Definition* left_def = curr_instr->left()->definition(); |
- Definition* right_def = curr_instr->right()->definition(); |
- for (intptr_t k = i + 1; k < merge_candidates->length(); k++) { |
- BinarySmiOpInstr* other_binop = (*merge_candidates)[k]; |
- // 'other_binop' can be NULL if it was already merged. |
- if ((other_binop != NULL) && |
- (other_binop->op_kind() == other_kind) && |
- (other_binop->left()->definition() == left_def) && |
- (other_binop->right()->definition() == right_def)) { |
- (*merge_candidates)[k] = NULL; // Clear it. |
- ASSERT(curr_instr->HasUses()); |
- AppendExtractNthOutputForMerged( |
- curr_instr, |
- MergedMathInstr::OutputIndexOf(curr_instr->op_kind()), |
- kTagged, kSmiCid); |
- ASSERT(other_binop->HasUses()); |
- AppendExtractNthOutputForMerged( |
- other_binop, |
- MergedMathInstr::OutputIndexOf(other_binop->op_kind()), |
- kTagged, kSmiCid); |
- |
- ZoneGrowableArray<Value*>* args = new(I) ZoneGrowableArray<Value*>(2); |
- args->Add(new(I) Value(curr_instr->left()->definition())); |
- args->Add(new(I) Value(curr_instr->right()->definition())); |
- |
- // Replace with TruncDivMod. |
- MergedMathInstr* div_mod = new(I) MergedMathInstr( |
- args, |
- curr_instr->deopt_id(), |
- MergedMathInstr::kTruncDivMod); |
- curr_instr->ReplaceWith(div_mod, current_iterator()); |
- other_binop->ReplaceUsesWith(div_mod); |
- other_binop->RemoveFromGraph(); |
- // Only one merge possible. Because canonicalization happens later, |
- // more candidates are possible. |
- // TODO(srdjan): Allow merging of trunc-div/mod into truncDivMod. |
- break; |
- } |
- } |
- } |
-} |
- |
- |
-// Tries to merge MathUnary operations, in this case sinus and cosinus. |
-void FlowGraphOptimizer::TryMergeMathUnary( |
- GrowableArray<MathUnaryInstr*>* merge_candidates) { |
- if (!FlowGraphCompiler::SupportsSinCos() || !CanUnboxDouble()) { |
- return; |
- } |
- if (merge_candidates->length() < 2) { |
- // Need at least a SIN and a COS. |
- return; |
- } |
- for (intptr_t i = 0; i < merge_candidates->length(); i++) { |
- MathUnaryInstr* curr_instr = (*merge_candidates)[i]; |
- if (curr_instr == NULL) { |
- // Instruction was merged already. |
- continue; |
- } |
- const intptr_t kind = curr_instr->kind(); |
- ASSERT((kind == MathUnaryInstr::kSin) || |
- (kind == MathUnaryInstr::kCos)); |
- // Check if there is sin/cos binop with same inputs. |
- const intptr_t other_kind = (kind == MethodRecognizer::kMathSin) ? |
- MethodRecognizer::kMathCos : MethodRecognizer::kMathSin; |
- Definition* def = curr_instr->value()->definition(); |
- for (intptr_t k = i + 1; k < merge_candidates->length(); k++) { |
- MathUnaryInstr* other_op = (*merge_candidates)[k]; |
- // 'other_op' can be NULL if it was already merged. |
- if ((other_op != NULL) && (other_op->kind() == other_kind) && |
- (other_op->value()->definition() == def)) { |
- (*merge_candidates)[k] = NULL; // Clear it. |
- ASSERT(curr_instr->HasUses()); |
- AppendExtractNthOutputForMerged(curr_instr, |
- MergedMathInstr::OutputIndexOf(kind), |
- kUnboxedDouble, kDoubleCid); |
- ASSERT(other_op->HasUses()); |
- AppendExtractNthOutputForMerged( |
- other_op, |
- MergedMathInstr::OutputIndexOf(other_kind), |
- kUnboxedDouble, kDoubleCid); |
- ZoneGrowableArray<Value*>* args = new(I) ZoneGrowableArray<Value*>(1); |
- args->Add(new(I) Value(curr_instr->value()->definition())); |
- // Replace with SinCos. |
- MergedMathInstr* sin_cos = |
- new(I) MergedMathInstr(args, |
- curr_instr->DeoptimizationTarget(), |
- MergedMathInstr::kSinCos); |
- curr_instr->ReplaceWith(sin_cos, current_iterator()); |
- other_op->ReplaceUsesWith(sin_cos); |
- other_op->RemoveFromGraph(); |
- // Only one merge possible. Because canonicalization happens later, |
- // more candidates are possible. |
- // TODO(srdjan): Allow merging of sin/cos into sincos. |
- break; |
- } |
- } |
- } |
-} |
- |
- |
-// Optimize (a << b) & c pattern: if c is a positive Smi or zero, then the |
-// shift can be a truncating Smi shift-left and result is always Smi. |
-// Merging occurs only per basic-block. |
-void FlowGraphOptimizer::TryOptimizePatterns() { |
- if (!FLAG_truncating_left_shift) return; |
- ASSERT(current_iterator_ == NULL); |
- GrowableArray<BinarySmiOpInstr*> div_mod_merge; |
- GrowableArray<MathUnaryInstr*> sin_cos_merge; |
- for (intptr_t i = 0; i < block_order_.length(); ++i) { |
- // Merging only per basic-block. |
- div_mod_merge.Clear(); |
- sin_cos_merge.Clear(); |
- BlockEntryInstr* entry = block_order_[i]; |
- ForwardInstructionIterator it(entry); |
- current_iterator_ = ⁢ |
- for (; !it.Done(); it.Advance()) { |
- if (it.Current()->IsBinarySmiOp()) { |
- BinarySmiOpInstr* binop = it.Current()->AsBinarySmiOp(); |
- if (binop->op_kind() == Token::kBIT_AND) { |
- OptimizeLeftShiftBitAndSmiOp(binop, |
- binop->left()->definition(), |
- binop->right()->definition()); |
- } else if ((binop->op_kind() == Token::kTRUNCDIV) || |
- (binop->op_kind() == Token::kMOD)) { |
- if (binop->HasUses()) { |
- div_mod_merge.Add(binop); |
- } |
- } |
- } else if (it.Current()->IsBinaryMintOp()) { |
- BinaryMintOpInstr* mintop = it.Current()->AsBinaryMintOp(); |
- if (mintop->op_kind() == Token::kBIT_AND) { |
- OptimizeLeftShiftBitAndSmiOp(mintop, |
- mintop->left()->definition(), |
- mintop->right()->definition()); |
- } |
- } else if (it.Current()->IsMathUnary()) { |
- MathUnaryInstr* math_unary = it.Current()->AsMathUnary(); |
- if ((math_unary->kind() == MathUnaryInstr::kSin) || |
- (math_unary->kind() == MathUnaryInstr::kCos)) { |
- if (math_unary->HasUses()) { |
- sin_cos_merge.Add(math_unary); |
- } |
- } |
- } |
- } |
- TryMergeTruncDivMod(&div_mod_merge); |
- TryMergeMathUnary(&sin_cos_merge); |
- current_iterator_ = NULL; |
- } |
-} |
- |
- |
-static void EnsureSSATempIndex(FlowGraph* graph, |
- Definition* defn, |
- Definition* replacement) { |
- if ((replacement->ssa_temp_index() == -1) && |
- (defn->ssa_temp_index() != -1)) { |
- replacement->set_ssa_temp_index(graph->alloc_ssa_temp_index()); |
- } |
-} |
- |
- |
-static void ReplaceCurrentInstruction(ForwardInstructionIterator* iterator, |
- Instruction* current, |
- Instruction* replacement, |
- FlowGraph* graph) { |
- Definition* current_defn = current->AsDefinition(); |
- if ((replacement != NULL) && (current_defn != NULL)) { |
- Definition* replacement_defn = replacement->AsDefinition(); |
- ASSERT(replacement_defn != NULL); |
- current_defn->ReplaceUsesWith(replacement_defn); |
- EnsureSSATempIndex(graph, current_defn, replacement_defn); |
- |
- if (FLAG_trace_optimization) { |
- OS::Print("Replacing v%" Pd " with v%" Pd "\n", |
- current_defn->ssa_temp_index(), |
- replacement_defn->ssa_temp_index()); |
- } |
- } else if (FLAG_trace_optimization) { |
- if (current_defn == NULL) { |
- OS::Print("Removing %s\n", current->DebugName()); |
- } else { |
- ASSERT(!current_defn->HasUses()); |
- OS::Print("Removing v%" Pd ".\n", current_defn->ssa_temp_index()); |
- } |
- } |
- iterator->RemoveCurrentFromGraph(); |
-} |
- |
- |
-bool FlowGraphOptimizer::Canonicalize() { |
- bool changed = false; |
- for (intptr_t i = 0; i < block_order_.length(); ++i) { |
- BlockEntryInstr* entry = block_order_[i]; |
- for (ForwardInstructionIterator it(entry); !it.Done(); it.Advance()) { |
- Instruction* current = it.Current(); |
- if (current->HasUnmatchedInputRepresentations()) { |
- // Can't canonicalize this instruction until all conversions for its |
- // inputs are inserted. |
- continue; |
- } |
- |
- Instruction* replacement = current->Canonicalize(flow_graph()); |
- |
- if (replacement != current) { |
- // For non-definitions Canonicalize should return either NULL or |
- // this. |
- ASSERT((replacement == NULL) || current->IsDefinition()); |
- ReplaceCurrentInstruction(&it, current, replacement, flow_graph_); |
- changed = true; |
- } |
- } |
- } |
- return changed; |
-} |
- |
- |
-static bool IsUnboxedInteger(Representation rep) { |
- return (rep == kUnboxedInt32) || |
- (rep == kUnboxedUint32) || |
- (rep == kUnboxedMint); |
-} |
- |
- |
-void FlowGraphOptimizer::InsertConversion(Representation from, |
- Representation to, |
- Value* use, |
- bool is_environment_use) { |
- Instruction* insert_before; |
- Instruction* deopt_target; |
- PhiInstr* phi = use->instruction()->AsPhi(); |
- if (phi != NULL) { |
- ASSERT(phi->is_alive()); |
- // For phis conversions have to be inserted in the predecessor. |
- insert_before = |
- phi->block()->PredecessorAt(use->use_index())->last_instruction(); |
- deopt_target = NULL; |
- } else { |
- deopt_target = insert_before = use->instruction(); |
- } |
- |
- Definition* converted = NULL; |
- if (IsUnboxedInteger(from) && IsUnboxedInteger(to)) { |
- const intptr_t deopt_id = (to == kUnboxedInt32) && (deopt_target != NULL) ? |
- deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId; |
- converted = new(I) UnboxedIntConverterInstr(from, |
- to, |
- use->CopyWithType(), |
- deopt_id); |
- } else if ((from == kUnboxedInt32) && (to == kUnboxedDouble)) { |
- converted = new Int32ToDoubleInstr(use->CopyWithType()); |
- } else if ((from == kUnboxedMint) && |
- (to == kUnboxedDouble) && |
- CanConvertUnboxedMintToDouble()) { |
- const intptr_t deopt_id = (deopt_target != NULL) ? |
- deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId; |
- ASSERT(CanUnboxDouble()); |
- converted = new MintToDoubleInstr(use->CopyWithType(), deopt_id); |
- } else if ((from == kTagged) && Boxing::Supports(to)) { |
- const intptr_t deopt_id = (deopt_target != NULL) ? |
- deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId; |
- converted = UnboxInstr::Create(to, use->CopyWithType(), deopt_id); |
- } else if ((to == kTagged) && Boxing::Supports(from)) { |
- converted = BoxInstr::Create(from, use->CopyWithType()); |
- } else { |
- // We have failed to find a suitable conversion instruction. |
- // Insert two "dummy" conversion instructions with the correct |
- // "from" and "to" representation. The inserted instructions will |
- // trigger a deoptimization if executed. See #12417 for a discussion. |
- const intptr_t deopt_id = (deopt_target != NULL) ? |
- deopt_target->DeoptimizationTarget() : Isolate::kNoDeoptId; |
- ASSERT(Boxing::Supports(from)); |
- ASSERT(Boxing::Supports(to)); |
- Definition* boxed = BoxInstr::Create(from, use->CopyWithType()); |
- use->BindTo(boxed); |
- InsertBefore(insert_before, boxed, NULL, FlowGraph::kValue); |
- converted = UnboxInstr::Create(to, new(I) Value(boxed), deopt_id); |
- } |
- ASSERT(converted != NULL); |
- InsertBefore(insert_before, converted, use->instruction()->env(), |
- FlowGraph::kValue); |
- if (is_environment_use) { |
- use->BindToEnvironment(converted); |
- } else { |
- use->BindTo(converted); |
- } |
- |
- if ((to == kUnboxedInt32) && (phi != NULL)) { |
- // Int32 phis are unboxed optimistically. Ensure that unboxing |
- // has deoptimization target attached from the goto instruction. |
- flow_graph_->CopyDeoptTarget(converted, insert_before); |
- } |
-} |
- |
- |
-void FlowGraphOptimizer::ConvertUse(Value* use, Representation from_rep) { |
- const Representation to_rep = |
- use->instruction()->RequiredInputRepresentation(use->use_index()); |
- if (from_rep == to_rep || to_rep == kNoRepresentation) { |
- return; |
- } |
- InsertConversion(from_rep, to_rep, use, /*is_environment_use=*/ false); |
-} |
- |
- |
-void FlowGraphOptimizer::ConvertEnvironmentUse(Value* use, |
- Representation from_rep) { |
- const Representation to_rep = kTagged; |
- if (from_rep == to_rep || to_rep == kNoRepresentation) { |
- return; |
- } |
- InsertConversion(from_rep, to_rep, use, /*is_environment_use=*/ true); |
-} |
- |
- |
-void FlowGraphOptimizer::InsertConversionsFor(Definition* def) { |
- const Representation from_rep = def->representation(); |
- |
- for (Value::Iterator it(def->input_use_list()); |
- !it.Done(); |
- it.Advance()) { |
- ConvertUse(it.Current(), from_rep); |
- } |
- |
- if (flow_graph()->graph_entry()->SuccessorCount() > 1) { |
- for (Value::Iterator it(def->env_use_list()); |
- !it.Done(); |
- it.Advance()) { |
- Value* use = it.Current(); |
- if (use->instruction()->MayThrow() && |
- use->instruction()->GetBlock()->InsideTryBlock()) { |
- // Environment uses at calls inside try-blocks must be converted to |
- // tagged representation. |
- ConvertEnvironmentUse(it.Current(), from_rep); |
- } |
- } |
- } |
-} |
- |
- |
-static void UnboxPhi(PhiInstr* phi) { |
- Representation unboxed = phi->representation(); |
- |
- switch (phi->Type()->ToCid()) { |
- case kDoubleCid: |
- if (CanUnboxDouble()) { |
- unboxed = kUnboxedDouble; |
- } |
- break; |
- case kFloat32x4Cid: |
- if (ShouldInlineSimd()) { |
- unboxed = kUnboxedFloat32x4; |
- } |
- break; |
- case kInt32x4Cid: |
- if (ShouldInlineSimd()) { |
- unboxed = kUnboxedInt32x4; |
- } |
- break; |
- case kFloat64x2Cid: |
- if (ShouldInlineSimd()) { |
- unboxed = kUnboxedFloat64x2; |
- } |
- break; |
- } |
- |
- if ((kSmiBits < 32) && |
- (unboxed == kTagged) && |
- phi->Type()->IsInt() && |
- RangeUtils::Fits(phi->range(), RangeBoundary::kRangeBoundaryInt32)) { |
- // On 32-bit platforms conservatively unbox phis that: |
- // - are proven to be of type Int; |
- // - fit into 32bits range; |
- // - have either constants or Box() operations as inputs; |
- // - have at least one Box() operation as an input; |
- // - are used in at least 1 Unbox() operation. |
- bool should_unbox = false; |
- for (intptr_t i = 0; i < phi->InputCount(); i++) { |
- Definition* input = phi->InputAt(i)->definition(); |
- if (input->IsBox() && |
- RangeUtils::Fits(input->range(), |
- RangeBoundary::kRangeBoundaryInt32)) { |
- should_unbox = true; |
- } else if (!input->IsConstant()) { |
- should_unbox = false; |
- break; |
- } |
- } |
- |
- if (should_unbox) { |
- // We checked inputs. Check if phi is used in at least one unbox |
- // operation. |
- bool has_unboxed_use = false; |
- for (Value* use = phi->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- Instruction* instr = use->instruction(); |
- if (instr->IsUnbox()) { |
- has_unboxed_use = true; |
- break; |
- } else if (IsUnboxedInteger( |
- instr->RequiredInputRepresentation(use->use_index()))) { |
- has_unboxed_use = true; |
- break; |
- } |
- } |
- |
- if (!has_unboxed_use) { |
- should_unbox = false; |
- } |
- } |
- |
- if (should_unbox) { |
- unboxed = kUnboxedInt32; |
- } |
- } |
- |
- phi->set_representation(unboxed); |
-} |
- |
- |
-void FlowGraphOptimizer::SelectRepresentations() { |
- // Conservatively unbox all phis that were proven to be of Double, |
- // Float32x4, or Int32x4 type. |
- for (intptr_t i = 0; i < block_order_.length(); ++i) { |
- JoinEntryInstr* join_entry = block_order_[i]->AsJoinEntry(); |
- if (join_entry != NULL) { |
- for (PhiIterator it(join_entry); !it.Done(); it.Advance()) { |
- PhiInstr* phi = it.Current(); |
- UnboxPhi(phi); |
- } |
- } |
- } |
- |
- // Process all instructions and insert conversions where needed. |
- GraphEntryInstr* graph_entry = block_order_[0]->AsGraphEntry(); |
- |
- // Visit incoming parameters and constants. |
- for (intptr_t i = 0; i < graph_entry->initial_definitions()->length(); i++) { |
- InsertConversionsFor((*graph_entry->initial_definitions())[i]); |
- } |
- |
- for (intptr_t i = 0; i < block_order_.length(); ++i) { |
- BlockEntryInstr* entry = block_order_[i]; |
- JoinEntryInstr* join_entry = entry->AsJoinEntry(); |
- if (join_entry != NULL) { |
- for (PhiIterator it(join_entry); !it.Done(); it.Advance()) { |
- PhiInstr* phi = it.Current(); |
- ASSERT(phi != NULL); |
- ASSERT(phi->is_alive()); |
- InsertConversionsFor(phi); |
- } |
- } |
- CatchBlockEntryInstr* catch_entry = entry->AsCatchBlockEntry(); |
- if (catch_entry != NULL) { |
- for (intptr_t i = 0; |
- i < catch_entry->initial_definitions()->length(); |
- i++) { |
- InsertConversionsFor((*catch_entry->initial_definitions())[i]); |
- } |
- } |
- for (ForwardInstructionIterator it(entry); !it.Done(); it.Advance()) { |
- Definition* def = it.Current()->AsDefinition(); |
- if (def != NULL) { |
- InsertConversionsFor(def); |
- } |
- } |
- } |
-} |
- |
- |
-static bool ClassIdIsOneOf(intptr_t class_id, |
- const GrowableArray<intptr_t>& class_ids) { |
- for (intptr_t i = 0; i < class_ids.length(); i++) { |
- ASSERT(class_ids[i] != kIllegalCid); |
- if (class_ids[i] == class_id) { |
- return true; |
- } |
- } |
- return false; |
-} |
- |
- |
-// Returns true if ICData tests two arguments and all ICData cids are in the |
-// required sets 'receiver_class_ids' or 'argument_class_ids', respectively. |
-static bool ICDataHasOnlyReceiverArgumentClassIds( |
- const ICData& ic_data, |
- const GrowableArray<intptr_t>& receiver_class_ids, |
- const GrowableArray<intptr_t>& argument_class_ids) { |
- if (ic_data.NumArgsTested() != 2) { |
- return false; |
- } |
- Function& target = Function::Handle(); |
- const intptr_t len = ic_data.NumberOfChecks(); |
- GrowableArray<intptr_t> class_ids; |
- for (intptr_t i = 0; i < len; i++) { |
- if (ic_data.IsUsedAt(i)) { |
- ic_data.GetCheckAt(i, &class_ids, &target); |
- ASSERT(class_ids.length() == 2); |
- if (!ClassIdIsOneOf(class_ids[0], receiver_class_ids) || |
- !ClassIdIsOneOf(class_ids[1], argument_class_ids)) { |
- return false; |
- } |
- } |
- } |
- return true; |
-} |
- |
- |
-static bool ICDataHasReceiverArgumentClassIds(const ICData& ic_data, |
- intptr_t receiver_class_id, |
- intptr_t argument_class_id) { |
- if (ic_data.NumArgsTested() != 2) { |
- return false; |
- } |
- Function& target = Function::Handle(); |
- const intptr_t len = ic_data.NumberOfChecks(); |
- for (intptr_t i = 0; i < len; i++) { |
- if (ic_data.IsUsedAt(i)) { |
- GrowableArray<intptr_t> class_ids; |
- ic_data.GetCheckAt(i, &class_ids, &target); |
- ASSERT(class_ids.length() == 2); |
- if ((class_ids[0] == receiver_class_id) && |
- (class_ids[1] == argument_class_id)) { |
- return true; |
- } |
- } |
- } |
- return false; |
-} |
- |
- |
-static bool HasOnlyOneSmi(const ICData& ic_data) { |
- return (ic_data.NumberOfUsedChecks() == 1) |
- && ic_data.HasReceiverClassId(kSmiCid); |
-} |
- |
- |
-static bool HasOnlySmiOrMint(const ICData& ic_data) { |
- if (ic_data.NumberOfUsedChecks() == 1) { |
- return ic_data.HasReceiverClassId(kSmiCid) |
- || ic_data.HasReceiverClassId(kMintCid); |
- } |
- return (ic_data.NumberOfUsedChecks() == 2) |
- && ic_data.HasReceiverClassId(kSmiCid) |
- && ic_data.HasReceiverClassId(kMintCid); |
-} |
- |
- |
-static bool HasOnlyTwoOf(const ICData& ic_data, intptr_t cid) { |
- if (ic_data.NumberOfUsedChecks() != 1) { |
- return false; |
- } |
- GrowableArray<intptr_t> first; |
- GrowableArray<intptr_t> second; |
- ic_data.GetUsedCidsForTwoArgs(&first, &second); |
- return (first[0] == cid) && (second[0] == cid); |
-} |
- |
-// Returns false if the ICData contains anything other than the 4 combinations |
-// of Mint and Smi for the receiver and argument classes. |
-static bool HasTwoMintOrSmi(const ICData& ic_data) { |
- GrowableArray<intptr_t> first; |
- GrowableArray<intptr_t> second; |
- ic_data.GetUsedCidsForTwoArgs(&first, &second); |
- for (intptr_t i = 0; i < first.length(); i++) { |
- if ((first[i] != kSmiCid) && (first[i] != kMintCid)) { |
- return false; |
- } |
- if ((second[i] != kSmiCid) && (second[i] != kMintCid)) { |
- return false; |
- } |
- } |
- return true; |
-} |
- |
- |
-// Returns false if the ICData contains anything other than the 4 combinations |
-// of Double and Smi for the receiver and argument classes. |
-static bool HasTwoDoubleOrSmi(const ICData& ic_data) { |
- GrowableArray<intptr_t> class_ids(2); |
- class_ids.Add(kSmiCid); |
- class_ids.Add(kDoubleCid); |
- return ICDataHasOnlyReceiverArgumentClassIds(ic_data, class_ids, class_ids); |
-} |
- |
- |
-static bool HasOnlyOneDouble(const ICData& ic_data) { |
- return (ic_data.NumberOfUsedChecks() == 1) |
- && ic_data.HasReceiverClassId(kDoubleCid); |
-} |
- |
- |
-static bool ShouldSpecializeForDouble(const ICData& ic_data) { |
- // Don't specialize for double if we can't unbox them. |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- |
- // Unboxed double operation can't handle case of two smis. |
- if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) { |
- return false; |
- } |
- |
- // Check that it have seen only smis and doubles. |
- return HasTwoDoubleOrSmi(ic_data); |
-} |
- |
- |
-void FlowGraphOptimizer::ReplaceCall(Definition* call, |
- Definition* replacement) { |
- // Remove the original push arguments. |
- for (intptr_t i = 0; i < call->ArgumentCount(); ++i) { |
- PushArgumentInstr* push = call->PushArgumentAt(i); |
- push->ReplaceUsesWith(push->value()->definition()); |
- push->RemoveFromGraph(); |
- } |
- call->ReplaceWith(replacement, current_iterator()); |
-} |
- |
- |
-void FlowGraphOptimizer::AddCheckSmi(Definition* to_check, |
- intptr_t deopt_id, |
- Environment* deopt_environment, |
- Instruction* insert_before) { |
- if (to_check->Type()->ToCid() != kSmiCid) { |
- InsertBefore(insert_before, |
- new(I) CheckSmiInstr(new(I) Value(to_check), |
- deopt_id, |
- insert_before->token_pos()), |
- deopt_environment, |
- FlowGraph::kEffect); |
- } |
-} |
- |
- |
-Instruction* FlowGraphOptimizer::GetCheckClass(Definition* to_check, |
- const ICData& unary_checks, |
- intptr_t deopt_id, |
- intptr_t token_pos) { |
- if ((unary_checks.NumberOfUsedChecks() == 1) && |
- unary_checks.HasReceiverClassId(kSmiCid)) { |
- return new(I) CheckSmiInstr(new(I) Value(to_check), |
- deopt_id, |
- token_pos); |
- } |
- return new(I) CheckClassInstr( |
- new(I) Value(to_check), deopt_id, unary_checks, token_pos); |
-} |
- |
- |
-void FlowGraphOptimizer::AddCheckClass(Definition* to_check, |
- const ICData& unary_checks, |
- intptr_t deopt_id, |
- Environment* deopt_environment, |
- Instruction* insert_before) { |
- // Type propagation has not run yet, we cannot eliminate the check. |
- Instruction* check = GetCheckClass( |
- to_check, unary_checks, deopt_id, insert_before->token_pos()); |
- InsertBefore(insert_before, check, deopt_environment, FlowGraph::kEffect); |
-} |
- |
- |
-void FlowGraphOptimizer::AddReceiverCheck(InstanceCallInstr* call) { |
- AddCheckClass(call->ArgumentAt(0), |
- ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks()), |
- call->deopt_id(), |
- call->env(), |
- call); |
-} |
- |
- |
-static bool ArgIsAlways(intptr_t cid, |
- const ICData& ic_data, |
- intptr_t arg_number) { |
- ASSERT(ic_data.NumArgsTested() > arg_number); |
- if (ic_data.NumberOfUsedChecks() == 0) { |
- return false; |
- } |
- const intptr_t num_checks = ic_data.NumberOfChecks(); |
- for (intptr_t i = 0; i < num_checks; i++) { |
- if (ic_data.IsUsedAt(i) && ic_data.GetClassIdAt(i, arg_number) != cid) { |
- return false; |
- } |
- } |
- return true; |
-} |
- |
- |
-static bool CanUnboxInt32() { |
- // Int32/Uint32 can be unboxed if it fits into a smi or the platform |
- // supports unboxed mints. |
- return (kSmiBits >= 32) || FlowGraphCompiler::SupportsUnboxedMints(); |
-} |
- |
- |
-static intptr_t MethodKindToCid(MethodRecognizer::Kind kind) { |
- switch (kind) { |
- case MethodRecognizer::kImmutableArrayGetIndexed: |
- return kImmutableArrayCid; |
- |
- case MethodRecognizer::kObjectArrayGetIndexed: |
- case MethodRecognizer::kObjectArraySetIndexed: |
- return kArrayCid; |
- |
- case MethodRecognizer::kGrowableArrayGetIndexed: |
- case MethodRecognizer::kGrowableArraySetIndexed: |
- return kGrowableObjectArrayCid; |
- |
- case MethodRecognizer::kFloat32ArrayGetIndexed: |
- case MethodRecognizer::kFloat32ArraySetIndexed: |
- return kTypedDataFloat32ArrayCid; |
- |
- case MethodRecognizer::kFloat64ArrayGetIndexed: |
- case MethodRecognizer::kFloat64ArraySetIndexed: |
- return kTypedDataFloat64ArrayCid; |
- |
- case MethodRecognizer::kInt8ArrayGetIndexed: |
- case MethodRecognizer::kInt8ArraySetIndexed: |
- return kTypedDataInt8ArrayCid; |
- |
- case MethodRecognizer::kUint8ArrayGetIndexed: |
- case MethodRecognizer::kUint8ArraySetIndexed: |
- return kTypedDataUint8ArrayCid; |
- |
- case MethodRecognizer::kUint8ClampedArrayGetIndexed: |
- case MethodRecognizer::kUint8ClampedArraySetIndexed: |
- return kTypedDataUint8ClampedArrayCid; |
- |
- case MethodRecognizer::kExternalUint8ArrayGetIndexed: |
- case MethodRecognizer::kExternalUint8ArraySetIndexed: |
- return kExternalTypedDataUint8ArrayCid; |
- |
- case MethodRecognizer::kExternalUint8ClampedArrayGetIndexed: |
- case MethodRecognizer::kExternalUint8ClampedArraySetIndexed: |
- return kExternalTypedDataUint8ClampedArrayCid; |
- |
- case MethodRecognizer::kInt16ArrayGetIndexed: |
- case MethodRecognizer::kInt16ArraySetIndexed: |
- return kTypedDataInt16ArrayCid; |
- |
- case MethodRecognizer::kUint16ArrayGetIndexed: |
- case MethodRecognizer::kUint16ArraySetIndexed: |
- return kTypedDataUint16ArrayCid; |
- |
- case MethodRecognizer::kInt32ArrayGetIndexed: |
- case MethodRecognizer::kInt32ArraySetIndexed: |
- return kTypedDataInt32ArrayCid; |
- |
- case MethodRecognizer::kUint32ArrayGetIndexed: |
- case MethodRecognizer::kUint32ArraySetIndexed: |
- return kTypedDataUint32ArrayCid; |
- |
- case MethodRecognizer::kInt64ArrayGetIndexed: |
- case MethodRecognizer::kInt64ArraySetIndexed: |
- return kTypedDataInt64ArrayCid; |
- |
- case MethodRecognizer::kFloat32x4ArrayGetIndexed: |
- case MethodRecognizer::kFloat32x4ArraySetIndexed: |
- return kTypedDataFloat32x4ArrayCid; |
- |
- case MethodRecognizer::kInt32x4ArrayGetIndexed: |
- case MethodRecognizer::kInt32x4ArraySetIndexed: |
- return kTypedDataInt32x4ArrayCid; |
- |
- case MethodRecognizer::kFloat64x2ArrayGetIndexed: |
- case MethodRecognizer::kFloat64x2ArraySetIndexed: |
- return kTypedDataFloat64x2ArrayCid; |
- |
- default: |
- break; |
- } |
- return kIllegalCid; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryReplaceWithStoreIndexed(InstanceCallInstr* call) { |
- // Check for monomorphic IC data. |
- if (!call->HasICData()) return false; |
- const ICData& ic_data = |
- ICData::Handle(I, call->ic_data()->AsUnaryClassChecks()); |
- if (ic_data.NumberOfChecks() != 1) { |
- return false; |
- } |
- ASSERT(ic_data.NumberOfUsedChecks() == 1); |
- ASSERT(ic_data.HasOneTarget()); |
- |
- const Function& target = Function::Handle(I, ic_data.GetTargetAt(0)); |
- TargetEntryInstr* entry; |
- Definition* last; |
- if (!TryInlineRecognizedMethod(ic_data.GetReceiverClassIdAt(0), |
- target, |
- call, |
- call->ArgumentAt(0), |
- call->token_pos(), |
- *call->ic_data(), |
- &entry, &last)) { |
- return false; |
- } |
- // Insert receiver class check. |
- AddReceiverCheck(call); |
- // Remove the original push arguments. |
- for (intptr_t i = 0; i < call->ArgumentCount(); ++i) { |
- PushArgumentInstr* push = call->PushArgumentAt(i); |
- push->ReplaceUsesWith(push->value()->definition()); |
- push->RemoveFromGraph(); |
- } |
- // Replace all uses of this definition with the result. |
- call->ReplaceUsesWith(last); |
- // Finally insert the sequence other definition in place of this one in the |
- // graph. |
- call->previous()->LinkTo(entry->next()); |
- entry->UnuseAllInputs(); // Entry block is not in the graph. |
- last->LinkTo(call); |
- // Remove through the iterator. |
- ASSERT(current_iterator()->Current() == call); |
- current_iterator()->RemoveCurrentFromGraph(); |
- call->set_previous(NULL); |
- call->set_next(NULL); |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineSetIndexed( |
- MethodRecognizer::Kind kind, |
- const Function& target, |
- Instruction* call, |
- Definition* receiver, |
- intptr_t token_pos, |
- const ICData* ic_data, |
- const ICData& value_check, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- intptr_t array_cid = MethodKindToCid(kind); |
- ASSERT(array_cid != kIllegalCid); |
- |
- Definition* array = receiver; |
- Definition* index = call->ArgumentAt(1); |
- Definition* stored_value = call->ArgumentAt(2); |
- |
- *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(), |
- call->GetBlock()->try_index()); |
- (*entry)->InheritDeoptTarget(I, call); |
- Instruction* cursor = *entry; |
- if (FLAG_enable_type_checks) { |
- // Only type check for the value. A type check for the index is not |
- // needed here because we insert a deoptimizing smi-check for the case |
- // the index is not a smi. |
- const AbstractType& value_type = |
- AbstractType::ZoneHandle(I, target.ParameterTypeAt(2)); |
- Definition* instantiator = NULL; |
- Definition* type_args = NULL; |
- switch (array_cid) { |
- case kArrayCid: |
- case kGrowableObjectArrayCid: { |
- const Class& instantiator_class = Class::Handle(I, target.Owner()); |
- intptr_t type_arguments_field_offset = |
- instantiator_class.type_arguments_field_offset(); |
- LoadFieldInstr* load_type_args = |
- new(I) LoadFieldInstr(new(I) Value(array), |
- type_arguments_field_offset, |
- Type::ZoneHandle(I), // No type. |
- call->token_pos()); |
- cursor = flow_graph()->AppendTo(cursor, |
- load_type_args, |
- NULL, |
- FlowGraph::kValue); |
- |
- instantiator = array; |
- type_args = load_type_args; |
- break; |
- } |
- case kTypedDataInt8ArrayCid: |
- case kTypedDataUint8ArrayCid: |
- case kTypedDataUint8ClampedArrayCid: |
- case kExternalTypedDataUint8ArrayCid: |
- case kExternalTypedDataUint8ClampedArrayCid: |
- case kTypedDataInt16ArrayCid: |
- case kTypedDataUint16ArrayCid: |
- case kTypedDataInt32ArrayCid: |
- case kTypedDataUint32ArrayCid: |
- case kTypedDataInt64ArrayCid: |
- ASSERT(value_type.IsIntType()); |
- // Fall through. |
- case kTypedDataFloat32ArrayCid: |
- case kTypedDataFloat64ArrayCid: { |
- type_args = instantiator = flow_graph_->constant_null(); |
- ASSERT((array_cid != kTypedDataFloat32ArrayCid && |
- array_cid != kTypedDataFloat64ArrayCid) || |
- value_type.IsDoubleType()); |
- ASSERT(value_type.IsInstantiated()); |
- break; |
- } |
- case kTypedDataFloat32x4ArrayCid: { |
- type_args = instantiator = flow_graph_->constant_null(); |
- ASSERT((array_cid != kTypedDataFloat32x4ArrayCid) || |
- value_type.IsFloat32x4Type()); |
- ASSERT(value_type.IsInstantiated()); |
- break; |
- } |
- case kTypedDataFloat64x2ArrayCid: { |
- type_args = instantiator = flow_graph_->constant_null(); |
- ASSERT((array_cid != kTypedDataFloat64x2ArrayCid) || |
- value_type.IsFloat64x2Type()); |
- ASSERT(value_type.IsInstantiated()); |
- break; |
- } |
- default: |
- // TODO(fschneider): Add support for other array types. |
- UNREACHABLE(); |
- } |
- AssertAssignableInstr* assert_value = |
- new(I) AssertAssignableInstr(token_pos, |
- new(I) Value(stored_value), |
- new(I) Value(instantiator), |
- new(I) Value(type_args), |
- value_type, |
- Symbols::Value(), |
- call->deopt_id()); |
- cursor = flow_graph()->AppendTo(cursor, |
- assert_value, |
- call->env(), |
- FlowGraph::kValue); |
- } |
- |
- array_cid = PrepareInlineIndexedOp(call, |
- array_cid, |
- &array, |
- index, |
- &cursor); |
- |
- // Check if store barrier is needed. Byte arrays don't need a store barrier. |
- StoreBarrierType needs_store_barrier = |
- (RawObject::IsTypedDataClassId(array_cid) || |
- RawObject::IsTypedDataViewClassId(array_cid) || |
- RawObject::IsExternalTypedDataClassId(array_cid)) ? kNoStoreBarrier |
- : kEmitStoreBarrier; |
- |
- // No need to class check stores to Int32 and Uint32 arrays because |
- // we insert unboxing instructions below which include a class check. |
- if ((array_cid != kTypedDataUint32ArrayCid) && |
- (array_cid != kTypedDataInt32ArrayCid) && |
- !value_check.IsNull()) { |
- // No store barrier needed because checked value is a smi, an unboxed mint, |
- // an unboxed double, an unboxed Float32x4, or unboxed Int32x4. |
- needs_store_barrier = kNoStoreBarrier; |
- Instruction* check = GetCheckClass( |
- stored_value, value_check, call->deopt_id(), call->token_pos()); |
- cursor = flow_graph()->AppendTo(cursor, |
- check, |
- call->env(), |
- FlowGraph::kEffect); |
- } |
- |
- if (array_cid == kTypedDataFloat32ArrayCid) { |
- stored_value = |
- new(I) DoubleToFloatInstr( |
- new(I) Value(stored_value), call->deopt_id()); |
- cursor = flow_graph()->AppendTo(cursor, |
- stored_value, |
- NULL, |
- FlowGraph::kValue); |
- } else if (array_cid == kTypedDataInt32ArrayCid) { |
- stored_value = new(I) UnboxInt32Instr( |
- UnboxInt32Instr::kTruncate, |
- new(I) Value(stored_value), |
- call->deopt_id()); |
- cursor = flow_graph()->AppendTo(cursor, |
- stored_value, |
- call->env(), |
- FlowGraph::kValue); |
- } else if (array_cid == kTypedDataUint32ArrayCid) { |
- stored_value = new(I) UnboxUint32Instr( |
- new(I) Value(stored_value), |
- call->deopt_id()); |
- ASSERT(stored_value->AsUnboxInteger()->is_truncating()); |
- cursor = flow_graph()->AppendTo(cursor, |
- stored_value, |
- call->env(), |
- FlowGraph::kValue); |
- } |
- |
- const intptr_t index_scale = Instance::ElementSizeFor(array_cid); |
- *last = new(I) StoreIndexedInstr(new(I) Value(array), |
- new(I) Value(index), |
- new(I) Value(stored_value), |
- needs_store_barrier, |
- index_scale, |
- array_cid, |
- call->deopt_id(), |
- call->token_pos()); |
- flow_graph()->AppendTo(cursor, |
- *last, |
- call->env(), |
- FlowGraph::kEffect); |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineRecognizedMethod(intptr_t receiver_cid, |
- const Function& target, |
- Instruction* call, |
- Definition* receiver, |
- intptr_t token_pos, |
- const ICData& ic_data, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- ICData& value_check = ICData::ZoneHandle(I); |
- MethodRecognizer::Kind kind = MethodRecognizer::RecognizeKind(target); |
- switch (kind) { |
- // Recognized [] operators. |
- case MethodRecognizer::kImmutableArrayGetIndexed: |
- case MethodRecognizer::kObjectArrayGetIndexed: |
- case MethodRecognizer::kGrowableArrayGetIndexed: |
- case MethodRecognizer::kInt8ArrayGetIndexed: |
- case MethodRecognizer::kUint8ArrayGetIndexed: |
- case MethodRecognizer::kUint8ClampedArrayGetIndexed: |
- case MethodRecognizer::kExternalUint8ArrayGetIndexed: |
- case MethodRecognizer::kExternalUint8ClampedArrayGetIndexed: |
- case MethodRecognizer::kInt16ArrayGetIndexed: |
- case MethodRecognizer::kUint16ArrayGetIndexed: |
- return InlineGetIndexed(kind, call, receiver, ic_data, entry, last); |
- case MethodRecognizer::kFloat32ArrayGetIndexed: |
- case MethodRecognizer::kFloat64ArrayGetIndexed: |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- return InlineGetIndexed(kind, call, receiver, ic_data, entry, last); |
- case MethodRecognizer::kFloat32x4ArrayGetIndexed: |
- case MethodRecognizer::kFloat64x2ArrayGetIndexed: |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- return InlineGetIndexed(kind, call, receiver, ic_data, entry, last); |
- case MethodRecognizer::kInt32ArrayGetIndexed: |
- case MethodRecognizer::kUint32ArrayGetIndexed: |
- if (!CanUnboxInt32()) return false; |
- return InlineGetIndexed(kind, call, receiver, ic_data, entry, last); |
- |
- case MethodRecognizer::kInt64ArrayGetIndexed: |
- if (!ShouldInlineInt64ArrayOps()) { |
- return false; |
- } |
- return InlineGetIndexed(kind, call, receiver, ic_data, entry, last); |
- // Recognized []= operators. |
- case MethodRecognizer::kObjectArraySetIndexed: |
- case MethodRecognizer::kGrowableArraySetIndexed: |
- if (ArgIsAlways(kSmiCid, ic_data, 2)) { |
- value_check = ic_data.AsUnaryClassChecksForArgNr(2); |
- } |
- return InlineSetIndexed(kind, target, call, receiver, token_pos, |
- &ic_data, value_check, entry, last); |
- case MethodRecognizer::kInt8ArraySetIndexed: |
- case MethodRecognizer::kUint8ArraySetIndexed: |
- case MethodRecognizer::kUint8ClampedArraySetIndexed: |
- case MethodRecognizer::kExternalUint8ArraySetIndexed: |
- case MethodRecognizer::kExternalUint8ClampedArraySetIndexed: |
- case MethodRecognizer::kInt16ArraySetIndexed: |
- case MethodRecognizer::kUint16ArraySetIndexed: |
- if (!ArgIsAlways(kSmiCid, ic_data, 2)) { |
- return false; |
- } |
- value_check = ic_data.AsUnaryClassChecksForArgNr(2); |
- return InlineSetIndexed(kind, target, call, receiver, token_pos, |
- &ic_data, value_check, entry, last); |
- case MethodRecognizer::kInt32ArraySetIndexed: |
- case MethodRecognizer::kUint32ArraySetIndexed: |
- // Check that value is always smi or mint. We use Int32/Uint32 unboxing |
- // which can only deal unbox these values. |
- value_check = ic_data.AsUnaryClassChecksForArgNr(2); |
- if (!HasOnlySmiOrMint(value_check)) { |
- return false; |
- } |
- return InlineSetIndexed(kind, target, call, receiver, token_pos, |
- &ic_data, value_check, entry, last); |
- case MethodRecognizer::kInt64ArraySetIndexed: |
- if (!ShouldInlineInt64ArrayOps()) { |
- return false; |
- } |
- return InlineSetIndexed(kind, target, call, receiver, token_pos, |
- &ic_data, value_check, entry, last); |
- case MethodRecognizer::kFloat32ArraySetIndexed: |
- case MethodRecognizer::kFloat64ArraySetIndexed: |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- // Check that value is always double. |
- if (!ArgIsAlways(kDoubleCid, ic_data, 2)) { |
- return false; |
- } |
- value_check = ic_data.AsUnaryClassChecksForArgNr(2); |
- return InlineSetIndexed(kind, target, call, receiver, token_pos, |
- &ic_data, value_check, entry, last); |
- case MethodRecognizer::kFloat32x4ArraySetIndexed: |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- // Check that value is always a Float32x4. |
- if (!ArgIsAlways(kFloat32x4Cid, ic_data, 2)) { |
- return false; |
- } |
- value_check = ic_data.AsUnaryClassChecksForArgNr(2); |
- return InlineSetIndexed(kind, target, call, receiver, token_pos, |
- &ic_data, value_check, entry, last); |
- case MethodRecognizer::kFloat64x2ArraySetIndexed: |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- // Check that value is always a Float32x4. |
- if (!ArgIsAlways(kFloat64x2Cid, ic_data, 2)) { |
- return false; |
- } |
- value_check = ic_data.AsUnaryClassChecksForArgNr(2); |
- return InlineSetIndexed(kind, target, call, receiver, token_pos, |
- &ic_data, value_check, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetInt8: |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataInt8ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetUint8: |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataUint8ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetInt16: |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataInt16ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetUint16: |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataUint16ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetInt32: |
- if (!CanUnboxInt32()) { |
- return false; |
- } |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataInt32ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetUint32: |
- if (!CanUnboxInt32()) { |
- return false; |
- } |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataUint32ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetFloat32: |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataFloat32ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetFloat64: |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataFloat64ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetFloat32x4: |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataFloat32x4ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseGetInt32x4: |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- return InlineByteArrayViewLoad(call, receiver, receiver_cid, |
- kTypedDataInt32x4ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetInt8: |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataInt8ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetUint8: |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataUint8ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetInt16: |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataInt16ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetUint16: |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataUint16ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetInt32: |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataInt32ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetUint32: |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataUint32ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetFloat32: |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataFloat32ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetFloat64: |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataFloat64ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetFloat32x4: |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataFloat32x4ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kByteArrayBaseSetInt32x4: |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- return InlineByteArrayViewStore(target, call, receiver, receiver_cid, |
- kTypedDataInt32x4ArrayCid, |
- ic_data, entry, last); |
- case MethodRecognizer::kStringBaseCodeUnitAt: |
- return InlineStringCodeUnitAt(call, receiver_cid, entry, last); |
- case MethodRecognizer::kStringBaseCharAt: |
- return InlineStringBaseCharAt(call, receiver_cid, entry, last); |
- case MethodRecognizer::kDoubleAdd: |
- return InlineDoubleOp(Token::kADD, call, entry, last); |
- case MethodRecognizer::kDoubleSub: |
- return InlineDoubleOp(Token::kSUB, call, entry, last); |
- case MethodRecognizer::kDoubleMul: |
- return InlineDoubleOp(Token::kMUL, call, entry, last); |
- case MethodRecognizer::kDoubleDiv: |
- return InlineDoubleOp(Token::kDIV, call, entry, last); |
- default: |
- return false; |
- } |
-} |
- |
- |
-intptr_t FlowGraphOptimizer::PrepareInlineIndexedOp(Instruction* call, |
- intptr_t array_cid, |
- Definition** array, |
- Definition* index, |
- Instruction** cursor) { |
- // Insert index smi check. |
- *cursor = flow_graph()->AppendTo( |
- *cursor, |
- new(I) CheckSmiInstr(new(I) Value(index), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- |
- // Insert array length load and bounds check. |
- LoadFieldInstr* length = |
- new(I) LoadFieldInstr( |
- new(I) Value(*array), |
- CheckArrayBoundInstr::LengthOffsetFor(array_cid), |
- Type::ZoneHandle(I, Type::SmiType()), |
- call->token_pos()); |
- length->set_is_immutable( |
- CheckArrayBoundInstr::IsFixedLengthArrayType(array_cid)); |
- length->set_result_cid(kSmiCid); |
- length->set_recognized_kind( |
- LoadFieldInstr::RecognizedKindFromArrayCid(array_cid)); |
- *cursor = flow_graph()->AppendTo(*cursor, |
- length, |
- NULL, |
- FlowGraph::kValue); |
- |
- *cursor = flow_graph()->AppendTo(*cursor, |
- new(I) CheckArrayBoundInstr( |
- new(I) Value(length), |
- new(I) Value(index), |
- call->deopt_id()), |
- call->env(), |
- FlowGraph::kEffect); |
- |
- if (array_cid == kGrowableObjectArrayCid) { |
- // Insert data elements load. |
- LoadFieldInstr* elements = |
- new(I) LoadFieldInstr( |
- new(I) Value(*array), |
- GrowableObjectArray::data_offset(), |
- Type::ZoneHandle(I, Type::DynamicType()), |
- call->token_pos()); |
- elements->set_result_cid(kArrayCid); |
- *cursor = flow_graph()->AppendTo(*cursor, |
- elements, |
- NULL, |
- FlowGraph::kValue); |
- // Load from the data from backing store which is a fixed-length array. |
- *array = elements; |
- array_cid = kArrayCid; |
- } else if (RawObject::IsExternalTypedDataClassId(array_cid)) { |
- LoadUntaggedInstr* elements = |
- new(I) LoadUntaggedInstr(new(I) Value(*array), |
- ExternalTypedData::data_offset()); |
- *cursor = flow_graph()->AppendTo(*cursor, |
- elements, |
- NULL, |
- FlowGraph::kValue); |
- *array = elements; |
- } |
- return array_cid; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineGetIndexed(MethodRecognizer::Kind kind, |
- Instruction* call, |
- Definition* receiver, |
- const ICData& ic_data, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- intptr_t array_cid = MethodKindToCid(kind); |
- ASSERT(array_cid != kIllegalCid); |
- |
- Definition* array = receiver; |
- Definition* index = call->ArgumentAt(1); |
- *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(), |
- call->GetBlock()->try_index()); |
- (*entry)->InheritDeoptTarget(I, call); |
- Instruction* cursor = *entry; |
- |
- array_cid = PrepareInlineIndexedOp(call, |
- array_cid, |
- &array, |
- index, |
- &cursor); |
- |
- intptr_t deopt_id = Isolate::kNoDeoptId; |
- if ((array_cid == kTypedDataInt32ArrayCid) || |
- (array_cid == kTypedDataUint32ArrayCid)) { |
- // Deoptimization may be needed if result does not always fit in a Smi. |
- deopt_id = (kSmiBits >= 32) ? Isolate::kNoDeoptId : call->deopt_id(); |
- } |
- |
- // Array load and return. |
- intptr_t index_scale = Instance::ElementSizeFor(array_cid); |
- *last = new(I) LoadIndexedInstr(new(I) Value(array), |
- new(I) Value(index), |
- index_scale, |
- array_cid, |
- deopt_id, |
- call->token_pos()); |
- cursor = flow_graph()->AppendTo( |
- cursor, |
- *last, |
- deopt_id != Isolate::kNoDeoptId ? call->env() : NULL, |
- FlowGraph::kValue); |
- |
- if (array_cid == kTypedDataFloat32ArrayCid) { |
- *last = new(I) FloatToDoubleInstr(new(I) Value(*last), deopt_id); |
- flow_graph()->AppendTo(cursor, |
- *last, |
- deopt_id != Isolate::kNoDeoptId ? call->env() : NULL, |
- FlowGraph::kValue); |
- } |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryReplaceWithLoadIndexed(InstanceCallInstr* call) { |
- // Check for monomorphic IC data. |
- if (!call->HasICData()) return false; |
- const ICData& ic_data = |
- ICData::Handle(I, call->ic_data()->AsUnaryClassChecks()); |
- if (ic_data.NumberOfChecks() != 1) { |
- return false; |
- } |
- ASSERT(ic_data.NumberOfUsedChecks() == 1); |
- ASSERT(ic_data.HasOneTarget()); |
- |
- const Function& target = Function::Handle(I, ic_data.GetTargetAt(0)); |
- TargetEntryInstr* entry; |
- Definition* last; |
- if (!TryInlineRecognizedMethod(ic_data.GetReceiverClassIdAt(0), |
- target, |
- call, |
- call->ArgumentAt(0), |
- call->token_pos(), |
- *call->ic_data(), |
- &entry, &last)) { |
- return false; |
- } |
- |
- // Insert receiver class check. |
- AddReceiverCheck(call); |
- // Remove the original push arguments. |
- for (intptr_t i = 0; i < call->ArgumentCount(); ++i) { |
- PushArgumentInstr* push = call->PushArgumentAt(i); |
- push->ReplaceUsesWith(push->value()->definition()); |
- push->RemoveFromGraph(); |
- } |
- // Replace all uses of this definition with the result. |
- call->ReplaceUsesWith(last); |
- // Finally insert the sequence other definition in place of this one in the |
- // graph. |
- call->previous()->LinkTo(entry->next()); |
- entry->UnuseAllInputs(); // Entry block is not in the graph. |
- last->LinkTo(call); |
- // Remove through the iterator. |
- ASSERT(current_iterator()->Current() == call); |
- current_iterator()->RemoveCurrentFromGraph(); |
- call->set_previous(NULL); |
- call->set_next(NULL); |
- return true; |
-} |
- |
- |
-// Return true if d is a string of length one (a constant or result from |
-// from string-from-char-code instruction. |
-static bool IsLengthOneString(Definition* d) { |
- if (d->IsConstant()) { |
- const Object& obj = d->AsConstant()->value(); |
- if (obj.IsString()) { |
- return String::Cast(obj).Length() == 1; |
- } else { |
- return false; |
- } |
- } else { |
- return d->IsStringFromCharCode(); |
- } |
-} |
- |
- |
-// Returns true if the string comparison was converted into char-code |
-// comparison. Conversion is only possible for strings of length one. |
-// E.g., detect str[x] == "x"; and use an integer comparison of char-codes. |
-// TODO(srdjan): Expand for two-byte and external strings. |
-bool FlowGraphOptimizer::TryStringLengthOneEquality(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- ASSERT(HasOnlyTwoOf(*call->ic_data(), kOneByteStringCid)); |
- // Check that left and right are length one strings (either string constants |
- // or results of string-from-char-code. |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- Value* left_val = NULL; |
- Definition* to_remove_left = NULL; |
- if (IsLengthOneString(right)) { |
- // Swap, since we know that both arguments are strings |
- Definition* temp = left; |
- left = right; |
- right = temp; |
- } |
- if (IsLengthOneString(left)) { |
- // Optimize if left is a string with length one (either constant or |
- // result of string-from-char-code. |
- if (left->IsConstant()) { |
- ConstantInstr* left_const = left->AsConstant(); |
- const String& str = String::Cast(left_const->value()); |
- ASSERT(str.Length() == 1); |
- ConstantInstr* char_code_left = flow_graph()->GetConstant( |
- Smi::ZoneHandle(I, Smi::New(static_cast<intptr_t>(str.CharAt(0))))); |
- left_val = new(I) Value(char_code_left); |
- } else if (left->IsStringFromCharCode()) { |
- // Use input of string-from-charcode as left value. |
- StringFromCharCodeInstr* instr = left->AsStringFromCharCode(); |
- left_val = new(I) Value(instr->char_code()->definition()); |
- to_remove_left = instr; |
- } else { |
- // IsLengthOneString(left) should have been false. |
- UNREACHABLE(); |
- } |
- |
- Definition* to_remove_right = NULL; |
- Value* right_val = NULL; |
- if (right->IsStringFromCharCode()) { |
- // Skip string-from-char-code, and use its input as right value. |
- StringFromCharCodeInstr* right_instr = right->AsStringFromCharCode(); |
- right_val = new(I) Value(right_instr->char_code()->definition()); |
- to_remove_right = right_instr; |
- } else { |
- const ICData& unary_checks_1 = |
- ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecksForArgNr(1)); |
- AddCheckClass(right, |
- unary_checks_1, |
- call->deopt_id(), |
- call->env(), |
- call); |
- // String-to-char-code instructions returns -1 (illegal charcode) if |
- // string is not of length one. |
- StringToCharCodeInstr* char_code_right = |
- new(I) StringToCharCodeInstr(new(I) Value(right), kOneByteStringCid); |
- InsertBefore(call, char_code_right, call->env(), FlowGraph::kValue); |
- right_val = new(I) Value(char_code_right); |
- } |
- |
- // Comparing char-codes instead of strings. |
- EqualityCompareInstr* comp = |
- new(I) EqualityCompareInstr(call->token_pos(), |
- op_kind, |
- left_val, |
- right_val, |
- kSmiCid, |
- call->deopt_id()); |
- ReplaceCall(call, comp); |
- |
- // Remove dead instructions. |
- if ((to_remove_left != NULL) && |
- (to_remove_left->input_use_list() == NULL)) { |
- to_remove_left->ReplaceUsesWith(flow_graph()->constant_null()); |
- to_remove_left->RemoveFromGraph(); |
- } |
- if ((to_remove_right != NULL) && |
- (to_remove_right->input_use_list() == NULL)) { |
- to_remove_right->ReplaceUsesWith(flow_graph()->constant_null()); |
- to_remove_right->RemoveFromGraph(); |
- } |
- return true; |
- } |
- return false; |
-} |
- |
- |
-static bool SmiFitsInDouble() { return kSmiBits < 53; } |
- |
-bool FlowGraphOptimizer::TryReplaceWithEqualityOp(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- const ICData& ic_data = *call->ic_data(); |
- ASSERT(ic_data.NumArgsTested() == 2); |
- |
- ASSERT(call->ArgumentCount() == 2); |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- |
- intptr_t cid = kIllegalCid; |
- if (HasOnlyTwoOf(ic_data, kOneByteStringCid)) { |
- if (TryStringLengthOneEquality(call, op_kind)) { |
- return true; |
- } else { |
- return false; |
- } |
- } else if (HasOnlyTwoOf(ic_data, kSmiCid)) { |
- InsertBefore(call, |
- new(I) CheckSmiInstr(new(I) Value(left), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- InsertBefore(call, |
- new(I) CheckSmiInstr(new(I) Value(right), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- cid = kSmiCid; |
- } else if (HasTwoMintOrSmi(ic_data) && |
- FlowGraphCompiler::SupportsUnboxedMints()) { |
- cid = kMintCid; |
- } else if (HasTwoDoubleOrSmi(ic_data) && CanUnboxDouble()) { |
- // Use double comparison. |
- if (SmiFitsInDouble()) { |
- cid = kDoubleCid; |
- } else { |
- if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) { |
- // We cannot use double comparison on two smis. Need polymorphic |
- // call. |
- return false; |
- } else { |
- InsertBefore(call, |
- new(I) CheckEitherNonSmiInstr( |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()), |
- call->env(), |
- FlowGraph::kEffect); |
- cid = kDoubleCid; |
- } |
- } |
- } else { |
- // Check if ICDData contains checks with Smi/Null combinations. In that case |
- // we can still emit the optimized Smi equality operation but need to add |
- // checks for null or Smi. |
- GrowableArray<intptr_t> smi_or_null(2); |
- smi_or_null.Add(kSmiCid); |
- smi_or_null.Add(kNullCid); |
- if (ICDataHasOnlyReceiverArgumentClassIds(ic_data, |
- smi_or_null, |
- smi_or_null)) { |
- const ICData& unary_checks_0 = |
- ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks()); |
- AddCheckClass(left, |
- unary_checks_0, |
- call->deopt_id(), |
- call->env(), |
- call); |
- |
- const ICData& unary_checks_1 = |
- ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecksForArgNr(1)); |
- AddCheckClass(right, |
- unary_checks_1, |
- call->deopt_id(), |
- call->env(), |
- call); |
- cid = kSmiCid; |
- } else { |
- // Shortcut for equality with null. |
- ConstantInstr* right_const = right->AsConstant(); |
- ConstantInstr* left_const = left->AsConstant(); |
- if ((right_const != NULL && right_const->value().IsNull()) || |
- (left_const != NULL && left_const->value().IsNull())) { |
- StrictCompareInstr* comp = |
- new(I) StrictCompareInstr(call->token_pos(), |
- Token::kEQ_STRICT, |
- new(I) Value(left), |
- new(I) Value(right), |
- false); // No number check. |
- ReplaceCall(call, comp); |
- return true; |
- } |
- return false; |
- } |
- } |
- ASSERT(cid != kIllegalCid); |
- EqualityCompareInstr* comp = new(I) EqualityCompareInstr(call->token_pos(), |
- op_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- cid, |
- call->deopt_id()); |
- ReplaceCall(call, comp); |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryReplaceWithRelationalOp(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- const ICData& ic_data = *call->ic_data(); |
- ASSERT(ic_data.NumArgsTested() == 2); |
- |
- ASSERT(call->ArgumentCount() == 2); |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- |
- intptr_t cid = kIllegalCid; |
- if (HasOnlyTwoOf(ic_data, kSmiCid)) { |
- InsertBefore(call, |
- new(I) CheckSmiInstr(new(I) Value(left), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- InsertBefore(call, |
- new(I) CheckSmiInstr(new(I) Value(right), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- cid = kSmiCid; |
- } else if (HasTwoMintOrSmi(ic_data) && |
- FlowGraphCompiler::SupportsUnboxedMints()) { |
- cid = kMintCid; |
- } else if (HasTwoDoubleOrSmi(ic_data) && CanUnboxDouble()) { |
- // Use double comparison. |
- if (SmiFitsInDouble()) { |
- cid = kDoubleCid; |
- } else { |
- if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) { |
- // We cannot use double comparison on two smis. Need polymorphic |
- // call. |
- return false; |
- } else { |
- InsertBefore(call, |
- new(I) CheckEitherNonSmiInstr( |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()), |
- call->env(), |
- FlowGraph::kEffect); |
- cid = kDoubleCid; |
- } |
- } |
- } else { |
- return false; |
- } |
- ASSERT(cid != kIllegalCid); |
- RelationalOpInstr* comp = new(I) RelationalOpInstr(call->token_pos(), |
- op_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- cid, |
- call->deopt_id()); |
- ReplaceCall(call, comp); |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryReplaceWithBinaryOp(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- intptr_t operands_type = kIllegalCid; |
- ASSERT(call->HasICData()); |
- const ICData& ic_data = *call->ic_data(); |
- switch (op_kind) { |
- case Token::kADD: |
- case Token::kSUB: |
- case Token::kMUL: |
- if (HasOnlyTwoOf(ic_data, kSmiCid)) { |
- // Don't generate smi code if the IC data is marked because |
- // of an overflow. |
- operands_type = ic_data.HasDeoptReason(ICData::kDeoptBinarySmiOp) |
- ? kMintCid |
- : kSmiCid; |
- } else if (HasTwoMintOrSmi(ic_data) && |
- FlowGraphCompiler::SupportsUnboxedMints()) { |
- // Don't generate mint code if the IC data is marked because of an |
- // overflow. |
- if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) return false; |
- operands_type = kMintCid; |
- } else if (ShouldSpecializeForDouble(ic_data)) { |
- operands_type = kDoubleCid; |
- } else if (HasOnlyTwoOf(ic_data, kFloat32x4Cid)) { |
- operands_type = kFloat32x4Cid; |
- } else if (HasOnlyTwoOf(ic_data, kInt32x4Cid)) { |
- ASSERT(op_kind != Token::kMUL); // Int32x4 doesn't have a multiply op. |
- operands_type = kInt32x4Cid; |
- } else if (HasOnlyTwoOf(ic_data, kFloat64x2Cid)) { |
- operands_type = kFloat64x2Cid; |
- } else { |
- return false; |
- } |
- break; |
- case Token::kDIV: |
- if (ShouldSpecializeForDouble(ic_data) || |
- HasOnlyTwoOf(ic_data, kSmiCid)) { |
- operands_type = kDoubleCid; |
- } else if (HasOnlyTwoOf(ic_data, kFloat32x4Cid)) { |
- operands_type = kFloat32x4Cid; |
- } else if (HasOnlyTwoOf(ic_data, kFloat64x2Cid)) { |
- operands_type = kFloat64x2Cid; |
- } else { |
- return false; |
- } |
- break; |
- case Token::kBIT_AND: |
- case Token::kBIT_OR: |
- case Token::kBIT_XOR: |
- if (HasOnlyTwoOf(ic_data, kSmiCid)) { |
- operands_type = kSmiCid; |
- } else if (HasTwoMintOrSmi(ic_data)) { |
- operands_type = kMintCid; |
- } else if (HasOnlyTwoOf(ic_data, kInt32x4Cid)) { |
- operands_type = kInt32x4Cid; |
- } else { |
- return false; |
- } |
- break; |
- case Token::kSHR: |
- case Token::kSHL: |
- if (HasOnlyTwoOf(ic_data, kSmiCid)) { |
- // Left shift may overflow from smi into mint or big ints. |
- // Don't generate smi code if the IC data is marked because |
- // of an overflow. |
- if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) { |
- return false; |
- } |
- operands_type = ic_data.HasDeoptReason(ICData::kDeoptBinarySmiOp) |
- ? kMintCid |
- : kSmiCid; |
- } else if (HasTwoMintOrSmi(ic_data) && |
- HasOnlyOneSmi(ICData::Handle(I, |
- ic_data.AsUnaryClassChecksForArgNr(1)))) { |
- // Don't generate mint code if the IC data is marked because of an |
- // overflow. |
- if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) { |
- return false; |
- } |
- // Check for smi/mint << smi or smi/mint >> smi. |
- operands_type = kMintCid; |
- } else { |
- return false; |
- } |
- break; |
- case Token::kMOD: |
- case Token::kTRUNCDIV: |
- if (HasOnlyTwoOf(ic_data, kSmiCid)) { |
- if (ic_data.HasDeoptReason(ICData::kDeoptBinarySmiOp)) { |
- return false; |
- } |
- operands_type = kSmiCid; |
- } else { |
- return false; |
- } |
- break; |
- default: |
- UNREACHABLE(); |
- } |
- |
- ASSERT(call->ArgumentCount() == 2); |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- if (operands_type == kDoubleCid) { |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- // Check that either left or right are not a smi. Result of a |
- // binary operation with two smis is a smi not a double, except '/' which |
- // returns a double for two smis. |
- if (op_kind != Token::kDIV) { |
- InsertBefore(call, |
- new(I) CheckEitherNonSmiInstr( |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()), |
- call->env(), |
- FlowGraph::kEffect); |
- } |
- |
- BinaryDoubleOpInstr* double_bin_op = |
- new(I) BinaryDoubleOpInstr(op_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id(), call->token_pos()); |
- ReplaceCall(call, double_bin_op); |
- } else if (operands_type == kMintCid) { |
- if (!FlowGraphCompiler::SupportsUnboxedMints()) return false; |
- if ((op_kind == Token::kSHR) || (op_kind == Token::kSHL)) { |
- ShiftMintOpInstr* shift_op = |
- new(I) ShiftMintOpInstr( |
- op_kind, new(I) Value(left), new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, shift_op); |
- } else { |
- BinaryMintOpInstr* bin_op = |
- new(I) BinaryMintOpInstr( |
- op_kind, new(I) Value(left), new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, bin_op); |
- } |
- } else if (operands_type == kFloat32x4Cid) { |
- return InlineFloat32x4BinaryOp(call, op_kind); |
- } else if (operands_type == kInt32x4Cid) { |
- return InlineInt32x4BinaryOp(call, op_kind); |
- } else if (operands_type == kFloat64x2Cid) { |
- return InlineFloat64x2BinaryOp(call, op_kind); |
- } else if (op_kind == Token::kMOD) { |
- ASSERT(operands_type == kSmiCid); |
- if (right->IsConstant()) { |
- const Object& obj = right->AsConstant()->value(); |
- if (obj.IsSmi() && Utils::IsPowerOfTwo(Smi::Cast(obj).Value())) { |
- // Insert smi check and attach a copy of the original environment |
- // because the smi operation can still deoptimize. |
- InsertBefore(call, |
- new(I) CheckSmiInstr(new(I) Value(left), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- ConstantInstr* constant = |
- flow_graph()->GetConstant(Smi::Handle(I, |
- Smi::New(Smi::Cast(obj).Value() - 1))); |
- BinarySmiOpInstr* bin_op = |
- new(I) BinarySmiOpInstr(Token::kBIT_AND, |
- new(I) Value(left), |
- new(I) Value(constant), |
- call->deopt_id()); |
- ReplaceCall(call, bin_op); |
- return true; |
- } |
- } |
- // Insert two smi checks and attach a copy of the original |
- // environment because the smi operation can still deoptimize. |
- AddCheckSmi(left, call->deopt_id(), call->env(), call); |
- AddCheckSmi(right, call->deopt_id(), call->env(), call); |
- BinarySmiOpInstr* bin_op = |
- new(I) BinarySmiOpInstr(op_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, bin_op); |
- } else { |
- ASSERT(operands_type == kSmiCid); |
- // Insert two smi checks and attach a copy of the original |
- // environment because the smi operation can still deoptimize. |
- AddCheckSmi(left, call->deopt_id(), call->env(), call); |
- AddCheckSmi(right, call->deopt_id(), call->env(), call); |
- if (left->IsConstant() && |
- ((op_kind == Token::kADD) || (op_kind == Token::kMUL))) { |
- // Constant should be on the right side. |
- Definition* temp = left; |
- left = right; |
- right = temp; |
- } |
- BinarySmiOpInstr* bin_op = |
- new(I) BinarySmiOpInstr( |
- op_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, bin_op); |
- } |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryReplaceWithUnaryOp(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- ASSERT(call->ArgumentCount() == 1); |
- Definition* input = call->ArgumentAt(0); |
- Definition* unary_op = NULL; |
- if (HasOnlyOneSmi(*call->ic_data())) { |
- InsertBefore(call, |
- new(I) CheckSmiInstr(new(I) Value(input), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- unary_op = new(I) UnarySmiOpInstr( |
- op_kind, new(I) Value(input), call->deopt_id()); |
- } else if ((op_kind == Token::kBIT_NOT) && |
- HasOnlySmiOrMint(*call->ic_data()) && |
- FlowGraphCompiler::SupportsUnboxedMints()) { |
- unary_op = new(I) UnaryMintOpInstr( |
- op_kind, new(I) Value(input), call->deopt_id()); |
- } else if (HasOnlyOneDouble(*call->ic_data()) && |
- (op_kind == Token::kNEGATE) && |
- CanUnboxDouble()) { |
- AddReceiverCheck(call); |
- unary_op = new(I) UnaryDoubleOpInstr( |
- Token::kNEGATE, new(I) Value(input), call->deopt_id()); |
- } else { |
- return false; |
- } |
- ASSERT(unary_op != NULL); |
- ReplaceCall(call, unary_op); |
- return true; |
-} |
- |
- |
-// Using field class |
-static RawField* GetField(intptr_t class_id, const String& field_name) { |
- Isolate* isolate = Isolate::Current(); |
- Class& cls = Class::Handle(isolate, isolate->class_table()->At(class_id)); |
- Field& field = Field::Handle(isolate); |
- while (!cls.IsNull()) { |
- field = cls.LookupInstanceField(field_name); |
- if (!field.IsNull()) { |
- return field.raw(); |
- } |
- cls = cls.SuperClass(); |
- } |
- return Field::null(); |
-} |
- |
- |
-// Use CHA to determine if the call needs a class check: if the callee's |
-// receiver is the same as the caller's receiver and there are no overriden |
-// callee functions, then no class check is needed. |
-bool FlowGraphOptimizer::InstanceCallNeedsClassCheck( |
- InstanceCallInstr* call, RawFunction::Kind kind) const { |
- if (!FLAG_use_cha) return true; |
- Definition* callee_receiver = call->ArgumentAt(0); |
- ASSERT(callee_receiver != NULL); |
- const Function& function = flow_graph_->parsed_function()->function(); |
- if (function.IsDynamicFunction() && |
- callee_receiver->IsParameter() && |
- (callee_receiver->AsParameter()->index() == 0)) { |
- const String& name = (kind == RawFunction::kMethodExtractor) |
- ? String::Handle(I, Field::NameFromGetter(call->function_name())) |
- : call->function_name(); |
- return isolate()->cha()->HasOverride(Class::Handle(I, function.Owner()), |
- name); |
- } |
- return true; |
-} |
- |
- |
-void FlowGraphOptimizer::InlineImplicitInstanceGetter(InstanceCallInstr* call) { |
- ASSERT(call->HasICData()); |
- const ICData& ic_data = *call->ic_data(); |
- ASSERT(ic_data.HasOneTarget()); |
- Function& target = Function::Handle(I); |
- GrowableArray<intptr_t> class_ids; |
- ic_data.GetCheckAt(0, &class_ids, &target); |
- ASSERT(class_ids.length() == 1); |
- // Inline implicit instance getter. |
- const String& field_name = |
- String::Handle(I, Field::NameFromGetter(call->function_name())); |
- const Field& field = |
- Field::ZoneHandle(I, GetField(class_ids[0], field_name)); |
- ASSERT(!field.IsNull()); |
- |
- if (InstanceCallNeedsClassCheck(call, RawFunction::kImplicitGetter)) { |
- AddReceiverCheck(call); |
- } |
- LoadFieldInstr* load = new(I) LoadFieldInstr( |
- new(I) Value(call->ArgumentAt(0)), |
- &field, |
- AbstractType::ZoneHandle(I, field.type()), |
- call->token_pos()); |
- load->set_is_immutable(field.is_final()); |
- if (field.guarded_cid() != kIllegalCid) { |
- if (!field.is_nullable() || (field.guarded_cid() == kNullCid)) { |
- load->set_result_cid(field.guarded_cid()); |
- } |
- FlowGraph::AddToGuardedFields(flow_graph_->guarded_fields(), &field); |
- } |
- |
- // Discard the environment from the original instruction because the load |
- // can't deoptimize. |
- call->RemoveEnvironment(); |
- ReplaceCall(call, load); |
- |
- if (load->result_cid() != kDynamicCid) { |
- // Reset value types if guarded_cid was used. |
- for (Value::Iterator it(load->input_use_list()); |
- !it.Done(); |
- it.Advance()) { |
- it.Current()->SetReachingType(NULL); |
- } |
- } |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineFloat32x4Getter(InstanceCallInstr* call, |
- MethodRecognizer::Kind getter) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- AddCheckClass(call->ArgumentAt(0), |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- intptr_t mask = 0; |
- if ((getter == MethodRecognizer::kFloat32x4Shuffle) || |
- (getter == MethodRecognizer::kFloat32x4ShuffleMix)) { |
- // Extract shuffle mask. |
- Definition* mask_definition = NULL; |
- if (getter == MethodRecognizer::kFloat32x4Shuffle) { |
- ASSERT(call->ArgumentCount() == 2); |
- mask_definition = call->ArgumentAt(1); |
- } else { |
- ASSERT(getter == MethodRecognizer::kFloat32x4ShuffleMix); |
- ASSERT(call->ArgumentCount() == 3); |
- mask_definition = call->ArgumentAt(2); |
- } |
- if (!mask_definition->IsConstant()) { |
- return false; |
- } |
- ASSERT(mask_definition->IsConstant()); |
- ConstantInstr* constant_instruction = mask_definition->AsConstant(); |
- const Object& constant_mask = constant_instruction->value(); |
- if (!constant_mask.IsSmi()) { |
- return false; |
- } |
- ASSERT(constant_mask.IsSmi()); |
- mask = Smi::Cast(constant_mask).Value(); |
- if ((mask < 0) || (mask > 255)) { |
- // Not a valid mask. |
- return false; |
- } |
- } |
- if (getter == MethodRecognizer::kFloat32x4GetSignMask) { |
- Simd32x4GetSignMaskInstr* instr = new(I) Simd32x4GetSignMaskInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } else if (getter == MethodRecognizer::kFloat32x4ShuffleMix) { |
- Simd32x4ShuffleMixInstr* instr = new(I) Simd32x4ShuffleMixInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- new(I) Value(call->ArgumentAt(1)), |
- mask, |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } else { |
- ASSERT((getter == MethodRecognizer::kFloat32x4Shuffle) || |
- (getter == MethodRecognizer::kFloat32x4ShuffleX) || |
- (getter == MethodRecognizer::kFloat32x4ShuffleY) || |
- (getter == MethodRecognizer::kFloat32x4ShuffleZ) || |
- (getter == MethodRecognizer::kFloat32x4ShuffleW)); |
- Simd32x4ShuffleInstr* instr = new(I) Simd32x4ShuffleInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- mask, |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } |
- UNREACHABLE(); |
- return false; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineFloat64x2Getter(InstanceCallInstr* call, |
- MethodRecognizer::Kind getter) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- AddCheckClass(call->ArgumentAt(0), |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- if ((getter == MethodRecognizer::kFloat64x2GetX) || |
- (getter == MethodRecognizer::kFloat64x2GetY)) { |
- Simd64x2ShuffleInstr* instr = new(I) Simd64x2ShuffleInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- 0, |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } |
- UNREACHABLE(); |
- return false; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineInt32x4Getter(InstanceCallInstr* call, |
- MethodRecognizer::Kind getter) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- AddCheckClass(call->ArgumentAt(0), |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- intptr_t mask = 0; |
- if ((getter == MethodRecognizer::kInt32x4Shuffle) || |
- (getter == MethodRecognizer::kInt32x4ShuffleMix)) { |
- // Extract shuffle mask. |
- Definition* mask_definition = NULL; |
- if (getter == MethodRecognizer::kInt32x4Shuffle) { |
- ASSERT(call->ArgumentCount() == 2); |
- mask_definition = call->ArgumentAt(1); |
- } else { |
- ASSERT(getter == MethodRecognizer::kInt32x4ShuffleMix); |
- ASSERT(call->ArgumentCount() == 3); |
- mask_definition = call->ArgumentAt(2); |
- } |
- if (!mask_definition->IsConstant()) { |
- return false; |
- } |
- ASSERT(mask_definition->IsConstant()); |
- ConstantInstr* constant_instruction = mask_definition->AsConstant(); |
- const Object& constant_mask = constant_instruction->value(); |
- if (!constant_mask.IsSmi()) { |
- return false; |
- } |
- ASSERT(constant_mask.IsSmi()); |
- mask = Smi::Cast(constant_mask).Value(); |
- if ((mask < 0) || (mask > 255)) { |
- // Not a valid mask. |
- return false; |
- } |
- } |
- if (getter == MethodRecognizer::kInt32x4GetSignMask) { |
- Simd32x4GetSignMaskInstr* instr = new(I) Simd32x4GetSignMaskInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } else if (getter == MethodRecognizer::kInt32x4ShuffleMix) { |
- Simd32x4ShuffleMixInstr* instr = new(I) Simd32x4ShuffleMixInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- new(I) Value(call->ArgumentAt(1)), |
- mask, |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } else if (getter == MethodRecognizer::kInt32x4Shuffle) { |
- Simd32x4ShuffleInstr* instr = new(I) Simd32x4ShuffleInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- mask, |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } else { |
- Int32x4GetFlagInstr* instr = new(I) Int32x4GetFlagInstr( |
- getter, |
- new(I) Value(call->ArgumentAt(0)), |
- call->deopt_id()); |
- ReplaceCall(call, instr); |
- return true; |
- } |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineFloat32x4BinaryOp(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- ASSERT(call->ArgumentCount() == 2); |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Type check right. |
- AddCheckClass(right, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(1)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Replace call. |
- BinaryFloat32x4OpInstr* float32x4_bin_op = |
- new(I) BinaryFloat32x4OpInstr( |
- op_kind, new(I) Value(left), new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, float32x4_bin_op); |
- |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineInt32x4BinaryOp(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- ASSERT(call->ArgumentCount() == 2); |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Type check right. |
- AddCheckClass(right, |
- ICData::ZoneHandle(I, |
- call->ic_data()->AsUnaryClassChecksForArgNr(1)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Replace call. |
- BinaryInt32x4OpInstr* int32x4_bin_op = |
- new(I) BinaryInt32x4OpInstr( |
- op_kind, new(I) Value(left), new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, int32x4_bin_op); |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineFloat64x2BinaryOp(InstanceCallInstr* call, |
- Token::Kind op_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- ASSERT(call->ArgumentCount() == 2); |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Type check right. |
- AddCheckClass(right, |
- ICData::ZoneHandle( |
- call->ic_data()->AsUnaryClassChecksForArgNr(1)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Replace call. |
- BinaryFloat64x2OpInstr* float64x2_bin_op = |
- new(I) BinaryFloat64x2OpInstr( |
- op_kind, new(I) Value(left), new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, float64x2_bin_op); |
- return true; |
-} |
- |
- |
-// Only unique implicit instance getters can be currently handled. |
-bool FlowGraphOptimizer::TryInlineInstanceGetter(InstanceCallInstr* call) { |
- ASSERT(call->HasICData()); |
- const ICData& ic_data = *call->ic_data(); |
- if (ic_data.NumberOfUsedChecks() == 0) { |
- // No type feedback collected. |
- return false; |
- } |
- |
- if (!ic_data.HasOneTarget()) { |
- // Polymorphic sites are inlined like normal methods by conventional |
- // inlining in FlowGraphInliner. |
- return false; |
- } |
- |
- const Function& target = Function::Handle(I, ic_data.GetTargetAt(0)); |
- if (target.kind() != RawFunction::kImplicitGetter) { |
- // Non-implicit getters are inlined like normal methods by conventional |
- // inlining in FlowGraphInliner. |
- return false; |
- } |
- InlineImplicitInstanceGetter(call); |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryReplaceInstanceCallWithInline( |
- InstanceCallInstr* call) { |
- ASSERT(call->HasICData()); |
- Function& target = Function::Handle(I); |
- GrowableArray<intptr_t> class_ids; |
- call->ic_data()->GetCheckAt(0, &class_ids, &target); |
- const intptr_t receiver_cid = class_ids[0]; |
- |
- TargetEntryInstr* entry; |
- Definition* last; |
- if (!TryInlineRecognizedMethod(receiver_cid, |
- target, |
- call, |
- call->ArgumentAt(0), |
- call->token_pos(), |
- *call->ic_data(), |
- &entry, &last)) { |
- return false; |
- } |
- |
- // Insert receiver class check. |
- AddReceiverCheck(call); |
- // Remove the original push arguments. |
- for (intptr_t i = 0; i < call->ArgumentCount(); ++i) { |
- PushArgumentInstr* push = call->PushArgumentAt(i); |
- push->ReplaceUsesWith(push->value()->definition()); |
- push->RemoveFromGraph(); |
- } |
- // Replace all uses of this definition with the result. |
- call->ReplaceUsesWith(last); |
- // Finally insert the sequence other definition in place of this one in the |
- // graph. |
- call->previous()->LinkTo(entry->next()); |
- entry->UnuseAllInputs(); // Entry block is not in the graph. |
- last->LinkTo(call); |
- // Remove through the iterator. |
- ASSERT(current_iterator()->Current() == call); |
- current_iterator()->RemoveCurrentFromGraph(); |
- call->set_previous(NULL); |
- call->set_next(NULL); |
- return true; |
-} |
- |
- |
-// Returns the LoadIndexedInstr. |
-Definition* FlowGraphOptimizer::PrepareInlineStringIndexOp( |
- Instruction* call, |
- intptr_t cid, |
- Definition* str, |
- Definition* index, |
- Instruction* cursor) { |
- |
- cursor = flow_graph()->AppendTo(cursor, |
- new(I) CheckSmiInstr( |
- new(I) Value(index), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- |
- // Load the length of the string. |
- // Treat length loads as mutable (i.e. affected by side effects) to avoid |
- // hoisting them since we can't hoist the preceding class-check. This |
- // is because of externalization of strings that affects their class-id. |
- LoadFieldInstr* length = new(I) LoadFieldInstr( |
- new(I) Value(str), |
- String::length_offset(), |
- Type::ZoneHandle(I, Type::SmiType()), |
- str->token_pos()); |
- length->set_result_cid(kSmiCid); |
- length->set_recognized_kind(MethodRecognizer::kStringBaseLength); |
- |
- cursor = flow_graph()->AppendTo(cursor, length, NULL, FlowGraph::kValue); |
- // Bounds check. |
- cursor = flow_graph()->AppendTo(cursor, |
- new(I) CheckArrayBoundInstr( |
- new(I) Value(length), |
- new(I) Value(index), |
- call->deopt_id()), |
- call->env(), |
- FlowGraph::kEffect); |
- |
- LoadIndexedInstr* load_indexed = new(I) LoadIndexedInstr( |
- new(I) Value(str), |
- new(I) Value(index), |
- Instance::ElementSizeFor(cid), |
- cid, |
- Isolate::kNoDeoptId, |
- call->token_pos()); |
- |
- cursor = flow_graph()->AppendTo(cursor, |
- load_indexed, |
- NULL, |
- FlowGraph::kValue); |
- ASSERT(cursor == load_indexed); |
- return load_indexed; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineStringCodeUnitAt( |
- Instruction* call, |
- intptr_t cid, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- // TODO(johnmccutchan): Handle external strings in PrepareInlineStringIndexOp. |
- if (RawObject::IsExternalStringClassId(cid)) { |
- return false; |
- } |
- |
- Definition* str = call->ArgumentAt(0); |
- Definition* index = call->ArgumentAt(1); |
- |
- *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(), |
- call->GetBlock()->try_index()); |
- (*entry)->InheritDeoptTarget(I, call); |
- |
- *last = PrepareInlineStringIndexOp(call, cid, str, index, *entry); |
- |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineStringBaseCharAt( |
- Instruction* call, |
- intptr_t cid, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- // TODO(johnmccutchan): Handle external strings in PrepareInlineStringIndexOp. |
- if (RawObject::IsExternalStringClassId(cid) || cid != kOneByteStringCid) { |
- return false; |
- } |
- Definition* str = call->ArgumentAt(0); |
- Definition* index = call->ArgumentAt(1); |
- |
- *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(), |
- call->GetBlock()->try_index()); |
- (*entry)->InheritDeoptTarget(I, call); |
- |
- *last = PrepareInlineStringIndexOp(call, cid, str, index, *entry); |
- |
- StringFromCharCodeInstr* char_at = new(I) StringFromCharCodeInstr( |
- new(I) Value(*last), cid); |
- |
- flow_graph()->AppendTo(*last, char_at, NULL, FlowGraph::kValue); |
- *last = char_at; |
- |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineDoubleOp( |
- Token::Kind op_kind, |
- Instruction* call, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- |
- *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(), |
- call->GetBlock()->try_index()); |
- (*entry)->InheritDeoptTarget(I, call); |
- // Arguments are checked. No need for class check. |
- BinaryDoubleOpInstr* double_bin_op = |
- new(I) BinaryDoubleOpInstr(op_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id(), call->token_pos()); |
- flow_graph()->AppendTo(*entry, double_bin_op, call->env(), FlowGraph::kValue); |
- *last = double_bin_op; |
- |
- return true; |
-} |
- |
- |
-void FlowGraphOptimizer::ReplaceWithMathCFunction( |
- InstanceCallInstr* call, |
- MethodRecognizer::Kind recognized_kind) { |
- AddReceiverCheck(call); |
- ZoneGrowableArray<Value*>* args = |
- new(I) ZoneGrowableArray<Value*>(call->ArgumentCount()); |
- for (intptr_t i = 0; i < call->ArgumentCount(); i++) { |
- args->Add(new(I) Value(call->ArgumentAt(i))); |
- } |
- InvokeMathCFunctionInstr* invoke = |
- new(I) InvokeMathCFunctionInstr(args, |
- call->deopt_id(), |
- recognized_kind, |
- call->token_pos()); |
- ReplaceCall(call, invoke); |
-} |
- |
- |
-static bool IsSupportedByteArrayViewCid(intptr_t cid) { |
- switch (cid) { |
- case kTypedDataInt8ArrayCid: |
- case kTypedDataUint8ArrayCid: |
- case kExternalTypedDataUint8ArrayCid: |
- case kTypedDataUint8ClampedArrayCid: |
- case kExternalTypedDataUint8ClampedArrayCid: |
- case kTypedDataInt16ArrayCid: |
- case kTypedDataUint16ArrayCid: |
- case kTypedDataInt32ArrayCid: |
- case kTypedDataUint32ArrayCid: |
- case kTypedDataFloat32ArrayCid: |
- case kTypedDataFloat64ArrayCid: |
- case kTypedDataFloat32x4ArrayCid: |
- case kTypedDataInt32x4ArrayCid: |
- return true; |
- default: |
- return false; |
- } |
-} |
- |
- |
-// Inline only simple, frequently called core library methods. |
-bool FlowGraphOptimizer::TryInlineInstanceMethod(InstanceCallInstr* call) { |
- ASSERT(call->HasICData()); |
- const ICData& ic_data = *call->ic_data(); |
- if ((ic_data.NumberOfUsedChecks() == 0) || !ic_data.HasOneTarget()) { |
- // No type feedback collected or multiple targets found. |
- return false; |
- } |
- |
- Function& target = Function::Handle(I); |
- GrowableArray<intptr_t> class_ids; |
- ic_data.GetCheckAt(0, &class_ids, &target); |
- MethodRecognizer::Kind recognized_kind = |
- MethodRecognizer::RecognizeKind(target); |
- |
- if ((recognized_kind == MethodRecognizer::kGrowableArraySetData) && |
- (ic_data.NumberOfChecks() == 1) && |
- (class_ids[0] == kGrowableObjectArrayCid)) { |
- // This is an internal method, no need to check argument types. |
- Definition* array = call->ArgumentAt(0); |
- Definition* value = call->ArgumentAt(1); |
- StoreInstanceFieldInstr* store = new(I) StoreInstanceFieldInstr( |
- GrowableObjectArray::data_offset(), |
- new(I) Value(array), |
- new(I) Value(value), |
- kEmitStoreBarrier, |
- call->token_pos()); |
- ReplaceCall(call, store); |
- return true; |
- } |
- |
- if ((recognized_kind == MethodRecognizer::kGrowableArraySetLength) && |
- (ic_data.NumberOfChecks() == 1) && |
- (class_ids[0] == kGrowableObjectArrayCid)) { |
- // This is an internal method, no need to check argument types nor |
- // range. |
- Definition* array = call->ArgumentAt(0); |
- Definition* value = call->ArgumentAt(1); |
- StoreInstanceFieldInstr* store = new(I) StoreInstanceFieldInstr( |
- GrowableObjectArray::length_offset(), |
- new(I) Value(array), |
- new(I) Value(value), |
- kNoStoreBarrier, |
- call->token_pos()); |
- ReplaceCall(call, store); |
- return true; |
- } |
- |
- if (((recognized_kind == MethodRecognizer::kStringBaseCodeUnitAt) || |
- (recognized_kind == MethodRecognizer::kStringBaseCharAt)) && |
- (ic_data.NumberOfChecks() == 1) && |
- ((class_ids[0] == kOneByteStringCid) || |
- (class_ids[0] == kTwoByteStringCid))) { |
- return TryReplaceInstanceCallWithInline(call); |
- } |
- |
- if ((class_ids[0] == kOneByteStringCid) && (ic_data.NumberOfChecks() == 1)) { |
- if (recognized_kind == MethodRecognizer::kOneByteStringSetAt) { |
- // This is an internal method, no need to check argument types nor |
- // range. |
- Definition* str = call->ArgumentAt(0); |
- Definition* index = call->ArgumentAt(1); |
- Definition* value = call->ArgumentAt(2); |
- StoreIndexedInstr* store_op = new(I) StoreIndexedInstr( |
- new(I) Value(str), |
- new(I) Value(index), |
- new(I) Value(value), |
- kNoStoreBarrier, |
- 1, // Index scale |
- kOneByteStringCid, |
- call->deopt_id(), |
- call->token_pos()); |
- ReplaceCall(call, store_op); |
- return true; |
- } |
- return false; |
- } |
- |
- if (CanUnboxDouble() && |
- (recognized_kind == MethodRecognizer::kIntegerToDouble) && |
- (ic_data.NumberOfChecks() == 1)) { |
- if (class_ids[0] == kSmiCid) { |
- AddReceiverCheck(call); |
- ReplaceCall(call, |
- new(I) SmiToDoubleInstr( |
- new(I) Value(call->ArgumentAt(0)), |
- call->token_pos())); |
- return true; |
- } else if ((class_ids[0] == kMintCid) && CanConvertUnboxedMintToDouble()) { |
- AddReceiverCheck(call); |
- ReplaceCall(call, |
- new(I) MintToDoubleInstr(new(I) Value(call->ArgumentAt(0)), |
- call->deopt_id())); |
- return true; |
- } |
- } |
- |
- if (class_ids[0] == kDoubleCid) { |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- switch (recognized_kind) { |
- case MethodRecognizer::kDoubleToInteger: { |
- AddReceiverCheck(call); |
- ASSERT(call->HasICData()); |
- const ICData& ic_data = *call->ic_data(); |
- Definition* input = call->ArgumentAt(0); |
- Definition* d2i_instr = NULL; |
- if (ic_data.HasDeoptReason(ICData::kDeoptDoubleToSmi)) { |
- // Do not repeatedly deoptimize because result didn't fit into Smi. |
- d2i_instr = new(I) DoubleToIntegerInstr( |
- new(I) Value(input), call); |
- } else { |
- // Optimistically assume result fits into Smi. |
- d2i_instr = new(I) DoubleToSmiInstr( |
- new(I) Value(input), call->deopt_id()); |
- } |
- ReplaceCall(call, d2i_instr); |
- return true; |
- } |
- case MethodRecognizer::kDoubleMod: |
- case MethodRecognizer::kDoubleRound: |
- ReplaceWithMathCFunction(call, recognized_kind); |
- return true; |
- case MethodRecognizer::kDoubleTruncate: |
- case MethodRecognizer::kDoubleFloor: |
- case MethodRecognizer::kDoubleCeil: |
- if (!TargetCPUFeatures::double_truncate_round_supported()) { |
- ReplaceWithMathCFunction(call, recognized_kind); |
- } else { |
- AddReceiverCheck(call); |
- DoubleToDoubleInstr* d2d_instr = |
- new(I) DoubleToDoubleInstr(new(I) Value(call->ArgumentAt(0)), |
- recognized_kind, call->deopt_id()); |
- ReplaceCall(call, d2d_instr); |
- } |
- return true; |
- case MethodRecognizer::kDoubleAdd: |
- case MethodRecognizer::kDoubleSub: |
- case MethodRecognizer::kDoubleMul: |
- case MethodRecognizer::kDoubleDiv: |
- return TryReplaceInstanceCallWithInline(call); |
- default: |
- // Unsupported method. |
- return false; |
- } |
- } |
- |
- if (IsSupportedByteArrayViewCid(class_ids[0]) && |
- (ic_data.NumberOfChecks() == 1)) { |
- // For elements that may not fit into a smi on all platforms, check if |
- // elements fit into a smi or the platform supports unboxed mints. |
- if ((recognized_kind == MethodRecognizer::kByteArrayBaseGetInt32) || |
- (recognized_kind == MethodRecognizer::kByteArrayBaseGetUint32) || |
- (recognized_kind == MethodRecognizer::kByteArrayBaseSetInt32) || |
- (recognized_kind == MethodRecognizer::kByteArrayBaseSetUint32)) { |
- if (!CanUnboxInt32()) { |
- return false; |
- } |
- } |
- |
- if ((recognized_kind == MethodRecognizer::kByteArrayBaseGetFloat32) || |
- (recognized_kind == MethodRecognizer::kByteArrayBaseGetFloat64) || |
- (recognized_kind == MethodRecognizer::kByteArrayBaseSetFloat32) || |
- (recognized_kind == MethodRecognizer::kByteArrayBaseSetFloat64)) { |
- if (!CanUnboxDouble()) { |
- return false; |
- } |
- } |
- |
- switch (recognized_kind) { |
- // ByteArray getters. |
- case MethodRecognizer::kByteArrayBaseGetInt8: |
- return BuildByteArrayViewLoad(call, kTypedDataInt8ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetUint8: |
- return BuildByteArrayViewLoad(call, kTypedDataUint8ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetInt16: |
- return BuildByteArrayViewLoad(call, kTypedDataInt16ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetUint16: |
- return BuildByteArrayViewLoad(call, kTypedDataUint16ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetInt32: |
- return BuildByteArrayViewLoad(call, kTypedDataInt32ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetUint32: |
- return BuildByteArrayViewLoad(call, kTypedDataUint32ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetFloat32: |
- return BuildByteArrayViewLoad(call, kTypedDataFloat32ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetFloat64: |
- return BuildByteArrayViewLoad(call, kTypedDataFloat64ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetFloat32x4: |
- return BuildByteArrayViewLoad(call, kTypedDataFloat32x4ArrayCid); |
- case MethodRecognizer::kByteArrayBaseGetInt32x4: |
- return BuildByteArrayViewLoad(call, kTypedDataInt32x4ArrayCid); |
- |
- // ByteArray setters. |
- case MethodRecognizer::kByteArrayBaseSetInt8: |
- return BuildByteArrayViewStore(call, kTypedDataInt8ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetUint8: |
- return BuildByteArrayViewStore(call, kTypedDataUint8ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetInt16: |
- return BuildByteArrayViewStore(call, kTypedDataInt16ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetUint16: |
- return BuildByteArrayViewStore(call, kTypedDataUint16ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetInt32: |
- return BuildByteArrayViewStore(call, kTypedDataInt32ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetUint32: |
- return BuildByteArrayViewStore(call, kTypedDataUint32ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetFloat32: |
- return BuildByteArrayViewStore(call, kTypedDataFloat32ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetFloat64: |
- return BuildByteArrayViewStore(call, kTypedDataFloat64ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetFloat32x4: |
- return BuildByteArrayViewStore(call, kTypedDataFloat32x4ArrayCid); |
- case MethodRecognizer::kByteArrayBaseSetInt32x4: |
- return BuildByteArrayViewStore(call, kTypedDataInt32x4ArrayCid); |
- default: |
- // Unsupported method. |
- return false; |
- } |
- } |
- |
- if ((class_ids[0] == kFloat32x4Cid) && (ic_data.NumberOfChecks() == 1)) { |
- return TryInlineFloat32x4Method(call, recognized_kind); |
- } |
- |
- if ((class_ids[0] == kInt32x4Cid) && (ic_data.NumberOfChecks() == 1)) { |
- return TryInlineInt32x4Method(call, recognized_kind); |
- } |
- |
- if ((class_ids[0] == kFloat64x2Cid) && (ic_data.NumberOfChecks() == 1)) { |
- return TryInlineFloat64x2Method(call, recognized_kind); |
- } |
- |
- if (recognized_kind == MethodRecognizer::kIntegerLeftShiftWithMask32) { |
- ASSERT(call->ArgumentCount() == 3); |
- ASSERT(ic_data.NumArgsTested() == 2); |
- Definition* value = call->ArgumentAt(0); |
- Definition* count = call->ArgumentAt(1); |
- Definition* int32_mask = call->ArgumentAt(2); |
- if (HasOnlyTwoOf(ic_data, kSmiCid)) { |
- if (ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) { |
- return false; |
- } |
- // We cannot overflow. The input value must be a Smi |
- AddCheckSmi(value, call->deopt_id(), call->env(), call); |
- AddCheckSmi(count, call->deopt_id(), call->env(), call); |
- ASSERT(int32_mask->IsConstant()); |
- const Integer& mask_literal = Integer::Cast( |
- int32_mask->AsConstant()->value()); |
- const int64_t mask_value = mask_literal.AsInt64Value(); |
- ASSERT(mask_value >= 0); |
- if (mask_value > Smi::kMaxValue) { |
- // The result will not be Smi. |
- return false; |
- } |
- BinarySmiOpInstr* left_shift = |
- new(I) BinarySmiOpInstr(Token::kSHL, |
- new(I) Value(value), |
- new(I) Value(count), |
- call->deopt_id()); |
- left_shift->mark_truncating(); |
- if ((kBitsPerWord == 32) && (mask_value == 0xffffffffLL)) { |
- // No BIT_AND operation needed. |
- ReplaceCall(call, left_shift); |
- } else { |
- InsertBefore(call, left_shift, call->env(), FlowGraph::kValue); |
- BinarySmiOpInstr* bit_and = |
- new(I) BinarySmiOpInstr(Token::kBIT_AND, |
- new(I) Value(left_shift), |
- new(I) Value(int32_mask), |
- call->deopt_id()); |
- ReplaceCall(call, bit_and); |
- } |
- return true; |
- } |
- |
- if (HasTwoMintOrSmi(ic_data) && |
- HasOnlyOneSmi(ICData::Handle(I, |
- ic_data.AsUnaryClassChecksForArgNr(1)))) { |
- if (!FlowGraphCompiler::SupportsUnboxedMints() || |
- ic_data.HasDeoptReason(ICData::kDeoptBinaryMintOp)) { |
- return false; |
- } |
- ShiftMintOpInstr* left_shift = |
- new(I) ShiftMintOpInstr(Token::kSHL, |
- new(I) Value(value), |
- new(I) Value(count), |
- call->deopt_id()); |
- InsertBefore(call, left_shift, call->env(), FlowGraph::kValue); |
- BinaryMintOpInstr* bit_and = |
- new(I) BinaryMintOpInstr(Token::kBIT_AND, |
- new(I) Value(left_shift), |
- new(I) Value(int32_mask), |
- call->deopt_id()); |
- ReplaceCall(call, bit_and); |
- return true; |
- } |
- } |
- return false; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineFloat32x4Constructor( |
- StaticCallInstr* call, |
- MethodRecognizer::Kind recognized_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- if (recognized_kind == MethodRecognizer::kFloat32x4Zero) { |
- Float32x4ZeroInstr* zero = new(I) Float32x4ZeroInstr(); |
- ReplaceCall(call, zero); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kFloat32x4Splat) { |
- Float32x4SplatInstr* splat = |
- new(I) Float32x4SplatInstr( |
- new(I) Value(call->ArgumentAt(1)), call->deopt_id()); |
- ReplaceCall(call, splat); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kFloat32x4Constructor) { |
- Float32x4ConstructorInstr* con = |
- new(I) Float32x4ConstructorInstr( |
- new(I) Value(call->ArgumentAt(1)), |
- new(I) Value(call->ArgumentAt(2)), |
- new(I) Value(call->ArgumentAt(3)), |
- new(I) Value(call->ArgumentAt(4)), |
- call->deopt_id()); |
- ReplaceCall(call, con); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kFloat32x4FromInt32x4Bits) { |
- Int32x4ToFloat32x4Instr* cast = |
- new(I) Int32x4ToFloat32x4Instr( |
- new(I) Value(call->ArgumentAt(1)), call->deopt_id()); |
- ReplaceCall(call, cast); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kFloat32x4FromFloat64x2) { |
- Float64x2ToFloat32x4Instr* cast = |
- new(I) Float64x2ToFloat32x4Instr( |
- new(I) Value(call->ArgumentAt(1)), call->deopt_id()); |
- ReplaceCall(call, cast); |
- return true; |
- } |
- return false; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineFloat64x2Constructor( |
- StaticCallInstr* call, |
- MethodRecognizer::Kind recognized_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- if (recognized_kind == MethodRecognizer::kFloat64x2Zero) { |
- Float64x2ZeroInstr* zero = new(I) Float64x2ZeroInstr(); |
- ReplaceCall(call, zero); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kFloat64x2Splat) { |
- Float64x2SplatInstr* splat = |
- new(I) Float64x2SplatInstr( |
- new(I) Value(call->ArgumentAt(1)), call->deopt_id()); |
- ReplaceCall(call, splat); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kFloat64x2Constructor) { |
- Float64x2ConstructorInstr* con = |
- new(I) Float64x2ConstructorInstr( |
- new(I) Value(call->ArgumentAt(1)), |
- new(I) Value(call->ArgumentAt(2)), |
- call->deopt_id()); |
- ReplaceCall(call, con); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kFloat64x2FromFloat32x4) { |
- Float32x4ToFloat64x2Instr* cast = |
- new(I) Float32x4ToFloat64x2Instr( |
- new(I) Value(call->ArgumentAt(1)), call->deopt_id()); |
- ReplaceCall(call, cast); |
- return true; |
- } |
- return false; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineInt32x4Constructor( |
- StaticCallInstr* call, |
- MethodRecognizer::Kind recognized_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- if (recognized_kind == MethodRecognizer::kInt32x4BoolConstructor) { |
- Int32x4BoolConstructorInstr* con = |
- new(I) Int32x4BoolConstructorInstr( |
- new(I) Value(call->ArgumentAt(1)), |
- new(I) Value(call->ArgumentAt(2)), |
- new(I) Value(call->ArgumentAt(3)), |
- new(I) Value(call->ArgumentAt(4)), |
- call->deopt_id()); |
- ReplaceCall(call, con); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kInt32x4FromFloat32x4Bits) { |
- Float32x4ToInt32x4Instr* cast = |
- new(I) Float32x4ToInt32x4Instr( |
- new(I) Value(call->ArgumentAt(1)), call->deopt_id()); |
- ReplaceCall(call, cast); |
- return true; |
- } else if (recognized_kind == MethodRecognizer::kInt32x4Constructor) { |
- Int32x4ConstructorInstr* con = |
- new(I) Int32x4ConstructorInstr( |
- new(I) Value(call->ArgumentAt(1)), |
- new(I) Value(call->ArgumentAt(2)), |
- new(I) Value(call->ArgumentAt(3)), |
- new(I) Value(call->ArgumentAt(4)), |
- call->deopt_id()); |
- ReplaceCall(call, con); |
- return true; |
- } |
- return false; |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineFloat32x4Method( |
- InstanceCallInstr* call, |
- MethodRecognizer::Kind recognized_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- ASSERT(call->HasICData()); |
- switch (recognized_kind) { |
- case MethodRecognizer::kFloat32x4ShuffleX: |
- case MethodRecognizer::kFloat32x4ShuffleY: |
- case MethodRecognizer::kFloat32x4ShuffleZ: |
- case MethodRecognizer::kFloat32x4ShuffleW: |
- case MethodRecognizer::kFloat32x4GetSignMask: |
- ASSERT(call->ic_data()->HasReceiverClassId(kFloat32x4Cid)); |
- ASSERT(call->ic_data()->HasOneTarget()); |
- return InlineFloat32x4Getter(call, recognized_kind); |
- |
- case MethodRecognizer::kFloat32x4Equal: |
- case MethodRecognizer::kFloat32x4GreaterThan: |
- case MethodRecognizer::kFloat32x4GreaterThanOrEqual: |
- case MethodRecognizer::kFloat32x4LessThan: |
- case MethodRecognizer::kFloat32x4LessThanOrEqual: |
- case MethodRecognizer::kFloat32x4NotEqual: { |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Replace call. |
- Float32x4ComparisonInstr* cmp = |
- new(I) Float32x4ComparisonInstr(recognized_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, cmp); |
- return true; |
- } |
- case MethodRecognizer::kFloat32x4Min: |
- case MethodRecognizer::kFloat32x4Max: { |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Float32x4MinMaxInstr* minmax = |
- new(I) Float32x4MinMaxInstr( |
- recognized_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, minmax); |
- return true; |
- } |
- case MethodRecognizer::kFloat32x4Scale: { |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- // Left and right values are swapped when handed to the instruction, |
- // this is done so that the double value is loaded into the output |
- // register and can be destroyed. |
- Float32x4ScaleInstr* scale = |
- new(I) Float32x4ScaleInstr(recognized_kind, |
- new(I) Value(right), |
- new(I) Value(left), |
- call->deopt_id()); |
- ReplaceCall(call, scale); |
- return true; |
- } |
- case MethodRecognizer::kFloat32x4Sqrt: |
- case MethodRecognizer::kFloat32x4ReciprocalSqrt: |
- case MethodRecognizer::kFloat32x4Reciprocal: { |
- Definition* left = call->ArgumentAt(0); |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Float32x4SqrtInstr* sqrt = |
- new(I) Float32x4SqrtInstr(recognized_kind, |
- new(I) Value(left), |
- call->deopt_id()); |
- ReplaceCall(call, sqrt); |
- return true; |
- } |
- case MethodRecognizer::kFloat32x4WithX: |
- case MethodRecognizer::kFloat32x4WithY: |
- case MethodRecognizer::kFloat32x4WithZ: |
- case MethodRecognizer::kFloat32x4WithW: { |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Float32x4WithInstr* with = new(I) Float32x4WithInstr(recognized_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, with); |
- return true; |
- } |
- case MethodRecognizer::kFloat32x4Absolute: |
- case MethodRecognizer::kFloat32x4Negate: { |
- Definition* left = call->ArgumentAt(0); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Float32x4ZeroArgInstr* zeroArg = |
- new(I) Float32x4ZeroArgInstr( |
- recognized_kind, new(I) Value(left), call->deopt_id()); |
- ReplaceCall(call, zeroArg); |
- return true; |
- } |
- case MethodRecognizer::kFloat32x4Clamp: { |
- Definition* left = call->ArgumentAt(0); |
- Definition* lower = call->ArgumentAt(1); |
- Definition* upper = call->ArgumentAt(2); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Float32x4ClampInstr* clamp = new(I) Float32x4ClampInstr( |
- new(I) Value(left), |
- new(I) Value(lower), |
- new(I) Value(upper), |
- call->deopt_id()); |
- ReplaceCall(call, clamp); |
- return true; |
- } |
- case MethodRecognizer::kFloat32x4ShuffleMix: |
- case MethodRecognizer::kFloat32x4Shuffle: { |
- return InlineFloat32x4Getter(call, recognized_kind); |
- } |
- default: |
- return false; |
- } |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineFloat64x2Method( |
- InstanceCallInstr* call, |
- MethodRecognizer::Kind recognized_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- ASSERT(call->HasICData()); |
- switch (recognized_kind) { |
- case MethodRecognizer::kFloat64x2GetX: |
- case MethodRecognizer::kFloat64x2GetY: |
- ASSERT(call->ic_data()->HasReceiverClassId(kFloat64x2Cid)); |
- ASSERT(call->ic_data()->HasOneTarget()); |
- return InlineFloat64x2Getter(call, recognized_kind); |
- case MethodRecognizer::kFloat64x2Negate: |
- case MethodRecognizer::kFloat64x2Abs: |
- case MethodRecognizer::kFloat64x2Sqrt: |
- case MethodRecognizer::kFloat64x2GetSignMask: { |
- Definition* left = call->ArgumentAt(0); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Float64x2ZeroArgInstr* zeroArg = |
- new(I) Float64x2ZeroArgInstr( |
- recognized_kind, new(I) Value(left), call->deopt_id()); |
- ReplaceCall(call, zeroArg); |
- return true; |
- } |
- case MethodRecognizer::kFloat64x2Scale: |
- case MethodRecognizer::kFloat64x2WithX: |
- case MethodRecognizer::kFloat64x2WithY: |
- case MethodRecognizer::kFloat64x2Min: |
- case MethodRecognizer::kFloat64x2Max: { |
- Definition* left = call->ArgumentAt(0); |
- Definition* right = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Float64x2OneArgInstr* zeroArg = |
- new(I) Float64x2OneArgInstr(recognized_kind, |
- new(I) Value(left), |
- new(I) Value(right), |
- call->deopt_id()); |
- ReplaceCall(call, zeroArg); |
- return true; |
- } |
- default: |
- return false; |
- } |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineInt32x4Method( |
- InstanceCallInstr* call, |
- MethodRecognizer::Kind recognized_kind) { |
- if (!ShouldInlineSimd()) { |
- return false; |
- } |
- ASSERT(call->HasICData()); |
- switch (recognized_kind) { |
- case MethodRecognizer::kInt32x4ShuffleMix: |
- case MethodRecognizer::kInt32x4Shuffle: |
- case MethodRecognizer::kInt32x4GetFlagX: |
- case MethodRecognizer::kInt32x4GetFlagY: |
- case MethodRecognizer::kInt32x4GetFlagZ: |
- case MethodRecognizer::kInt32x4GetFlagW: |
- case MethodRecognizer::kInt32x4GetSignMask: |
- ASSERT(call->ic_data()->HasReceiverClassId(kInt32x4Cid)); |
- ASSERT(call->ic_data()->HasOneTarget()); |
- return InlineInt32x4Getter(call, recognized_kind); |
- |
- case MethodRecognizer::kInt32x4Select: { |
- Definition* mask = call->ArgumentAt(0); |
- Definition* trueValue = call->ArgumentAt(1); |
- Definition* falseValue = call->ArgumentAt(2); |
- // Type check left. |
- AddCheckClass(mask, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Int32x4SelectInstr* select = new(I) Int32x4SelectInstr( |
- new(I) Value(mask), |
- new(I) Value(trueValue), |
- new(I) Value(falseValue), |
- call->deopt_id()); |
- ReplaceCall(call, select); |
- return true; |
- } |
- case MethodRecognizer::kInt32x4WithFlagX: |
- case MethodRecognizer::kInt32x4WithFlagY: |
- case MethodRecognizer::kInt32x4WithFlagZ: |
- case MethodRecognizer::kInt32x4WithFlagW: { |
- Definition* left = call->ArgumentAt(0); |
- Definition* flag = call->ArgumentAt(1); |
- // Type check left. |
- AddCheckClass(left, |
- ICData::ZoneHandle( |
- I, call->ic_data()->AsUnaryClassChecksForArgNr(0)), |
- call->deopt_id(), |
- call->env(), |
- call); |
- Int32x4SetFlagInstr* setFlag = new(I) Int32x4SetFlagInstr( |
- recognized_kind, |
- new(I) Value(left), |
- new(I) Value(flag), |
- call->deopt_id()); |
- ReplaceCall(call, setFlag); |
- return true; |
- } |
- default: |
- return false; |
- } |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineByteArrayViewLoad(Instruction* call, |
- Definition* receiver, |
- intptr_t array_cid, |
- intptr_t view_cid, |
- const ICData& ic_data, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- ASSERT(array_cid != kIllegalCid); |
- Definition* array = receiver; |
- Definition* index = call->ArgumentAt(1); |
- *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(), |
- call->GetBlock()->try_index()); |
- (*entry)->InheritDeoptTarget(I, call); |
- Instruction* cursor = *entry; |
- |
- array_cid = PrepareInlineByteArrayViewOp(call, |
- array_cid, |
- view_cid, |
- &array, |
- index, |
- &cursor); |
- |
- intptr_t deopt_id = Isolate::kNoDeoptId; |
- if ((array_cid == kTypedDataInt32ArrayCid) || |
- (array_cid == kTypedDataUint32ArrayCid)) { |
- // Deoptimization may be needed if result does not always fit in a Smi. |
- deopt_id = (kSmiBits >= 32) ? Isolate::kNoDeoptId : call->deopt_id(); |
- } |
- |
- *last = new(I) LoadIndexedInstr(new(I) Value(array), |
- new(I) Value(index), |
- 1, |
- view_cid, |
- deopt_id, |
- call->token_pos()); |
- cursor = flow_graph()->AppendTo( |
- cursor, |
- *last, |
- deopt_id != Isolate::kNoDeoptId ? call->env() : NULL, |
- FlowGraph::kValue); |
- |
- if (view_cid == kTypedDataFloat32ArrayCid) { |
- *last = new(I) FloatToDoubleInstr(new(I) Value(*last), deopt_id); |
- flow_graph()->AppendTo(cursor, |
- *last, |
- deopt_id != Isolate::kNoDeoptId ? call->env() : NULL, |
- FlowGraph::kValue); |
- } |
- return true; |
-} |
- |
- |
-bool FlowGraphOptimizer::InlineByteArrayViewStore(const Function& target, |
- Instruction* call, |
- Definition* receiver, |
- intptr_t array_cid, |
- intptr_t view_cid, |
- const ICData& ic_data, |
- TargetEntryInstr** entry, |
- Definition** last) { |
- ASSERT(array_cid != kIllegalCid); |
- Definition* array = receiver; |
- Definition* index = call->ArgumentAt(1); |
- *entry = new(I) TargetEntryInstr(flow_graph()->allocate_block_id(), |
- call->GetBlock()->try_index()); |
- (*entry)->InheritDeoptTarget(I, call); |
- Instruction* cursor = *entry; |
- |
- array_cid = PrepareInlineByteArrayViewOp(call, |
- array_cid, |
- view_cid, |
- &array, |
- index, |
- &cursor); |
- |
- // Extract the instance call so we can use the function_name in the stored |
- // value check ICData. |
- InstanceCallInstr* i_call = NULL; |
- if (call->IsPolymorphicInstanceCall()) { |
- i_call = call->AsPolymorphicInstanceCall()->instance_call(); |
- } else { |
- ASSERT(call->IsInstanceCall()); |
- i_call = call->AsInstanceCall(); |
- } |
- ASSERT(i_call != NULL); |
- ICData& value_check = ICData::ZoneHandle(I); |
- switch (view_cid) { |
- case kTypedDataInt8ArrayCid: |
- case kTypedDataUint8ArrayCid: |
- case kTypedDataUint8ClampedArrayCid: |
- case kExternalTypedDataUint8ArrayCid: |
- case kExternalTypedDataUint8ClampedArrayCid: |
- case kTypedDataInt16ArrayCid: |
- case kTypedDataUint16ArrayCid: { |
- // Check that value is always smi. |
- value_check = ICData::New(flow_graph_->parsed_function()->function(), |
- i_call->function_name(), |
- Object::empty_array(), // Dummy args. descr. |
- Isolate::kNoDeoptId, |
- 1); |
- value_check.AddReceiverCheck(kSmiCid, target); |
- break; |
- } |
- case kTypedDataInt32ArrayCid: |
- case kTypedDataUint32ArrayCid: |
- // On 64-bit platforms assume that stored value is always a smi. |
- if (kSmiBits >= 32) { |
- value_check = ICData::New(flow_graph_->parsed_function()->function(), |
- i_call->function_name(), |
- Object::empty_array(), // Dummy args. descr. |
- Isolate::kNoDeoptId, |
- 1); |
- value_check.AddReceiverCheck(kSmiCid, target); |
- } |
- break; |
- case kTypedDataFloat32ArrayCid: |
- case kTypedDataFloat64ArrayCid: { |
- // Check that value is always double. |
- value_check = ICData::New(flow_graph_->parsed_function()->function(), |
- i_call->function_name(), |
- Object::empty_array(), // Dummy args. descr. |
- Isolate::kNoDeoptId, |
- 1); |
- value_check.AddReceiverCheck(kDoubleCid, target); |
- break; |
- } |
- case kTypedDataInt32x4ArrayCid: { |
- // Check that value is always Int32x4. |
- value_check = ICData::New(flow_graph_->parsed_function()->function(), |
- i_call->function_name(), |
- Object::empty_array(), // Dummy args. descr. |
- Isolate::kNoDeoptId, |
- 1); |
- value_check.AddReceiverCheck(kInt32x4Cid, target); |
- break; |
- } |
- case kTypedDataFloat32x4ArrayCid: { |
- // Check that value is always Float32x4. |
- value_check = ICData::New(flow_graph_->parsed_function()->function(), |
- i_call->function_name(), |
- Object::empty_array(), // Dummy args. descr. |
- Isolate::kNoDeoptId, |
- 1); |
- value_check.AddReceiverCheck(kFloat32x4Cid, target); |
- break; |
- } |
- default: |
- // Array cids are already checked in the caller. |
- UNREACHABLE(); |
- } |
- |
- Definition* stored_value = call->ArgumentAt(2); |
- if (!value_check.IsNull()) { |
- AddCheckClass(stored_value, value_check, call->deopt_id(), call->env(), |
- call); |
- } |
- |
- if (view_cid == kTypedDataFloat32ArrayCid) { |
- stored_value = new(I) DoubleToFloatInstr( |
- new(I) Value(stored_value), call->deopt_id()); |
- cursor = flow_graph()->AppendTo(cursor, |
- stored_value, |
- NULL, |
- FlowGraph::kValue); |
- } else if (view_cid == kTypedDataInt32ArrayCid) { |
- stored_value = new(I) UnboxInt32Instr( |
- UnboxInt32Instr::kTruncate, |
- new(I) Value(stored_value), |
- call->deopt_id()); |
- cursor = flow_graph()->AppendTo(cursor, |
- stored_value, |
- call->env(), |
- FlowGraph::kValue); |
- } else if (view_cid == kTypedDataUint32ArrayCid) { |
- stored_value = new(I) UnboxUint32Instr( |
- new(I) Value(stored_value), |
- call->deopt_id()); |
- ASSERT(stored_value->AsUnboxInteger()->is_truncating()); |
- cursor = flow_graph()->AppendTo(cursor, |
- stored_value, |
- call->env(), |
- FlowGraph::kValue); |
- } |
- |
- StoreBarrierType needs_store_barrier = kNoStoreBarrier; |
- *last = new(I) StoreIndexedInstr(new(I) Value(array), |
- new(I) Value(index), |
- new(I) Value(stored_value), |
- needs_store_barrier, |
- 1, // Index scale |
- view_cid, |
- call->deopt_id(), |
- call->token_pos()); |
- |
- flow_graph()->AppendTo(cursor, |
- *last, |
- call->deopt_id() != Isolate::kNoDeoptId ? |
- call->env() : NULL, |
- FlowGraph::kEffect); |
- return true; |
-} |
- |
- |
- |
-intptr_t FlowGraphOptimizer::PrepareInlineByteArrayViewOp( |
- Instruction* call, |
- intptr_t array_cid, |
- intptr_t view_cid, |
- Definition** array, |
- Definition* byte_index, |
- Instruction** cursor) { |
- // Insert byte_index smi check. |
- *cursor = flow_graph()->AppendTo(*cursor, |
- new(I) CheckSmiInstr( |
- new(I) Value(byte_index), |
- call->deopt_id(), |
- call->token_pos()), |
- call->env(), |
- FlowGraph::kEffect); |
- |
- LoadFieldInstr* length = |
- new(I) LoadFieldInstr( |
- new(I) Value(*array), |
- CheckArrayBoundInstr::LengthOffsetFor(array_cid), |
- Type::ZoneHandle(I, Type::SmiType()), |
- call->token_pos()); |
- length->set_is_immutable(true); |
- length->set_result_cid(kSmiCid); |
- length->set_recognized_kind( |
- LoadFieldInstr::RecognizedKindFromArrayCid(array_cid)); |
- *cursor = flow_graph()->AppendTo(*cursor, |
- length, |
- NULL, |
- FlowGraph::kValue); |
- |
- intptr_t element_size = Instance::ElementSizeFor(array_cid); |
- ConstantInstr* bytes_per_element = |
- flow_graph()->GetConstant(Smi::Handle(I, Smi::New(element_size))); |
- BinarySmiOpInstr* len_in_bytes = |
- new(I) BinarySmiOpInstr(Token::kMUL, |
- new(I) Value(length), |
- new(I) Value(bytes_per_element), |
- call->deopt_id()); |
- *cursor = flow_graph()->AppendTo(*cursor, len_in_bytes, call->env(), |
- FlowGraph::kValue); |
- |
- // adjusted_length = len_in_bytes - (element_size - 1). |
- Definition* adjusted_length = len_in_bytes; |
- intptr_t adjustment = Instance::ElementSizeFor(view_cid) - 1; |
- if (adjustment > 0) { |
- ConstantInstr* length_adjustment = |
- flow_graph()->GetConstant(Smi::Handle(I, Smi::New(adjustment))); |
- adjusted_length = |
- new(I) BinarySmiOpInstr(Token::kSUB, |
- new(I) Value(len_in_bytes), |
- new(I) Value(length_adjustment), |
- call->deopt_id()); |
- *cursor = flow_graph()->AppendTo(*cursor, adjusted_length, call->env(), |
- FlowGraph::kValue); |
- } |
- |
- // Check adjusted_length > 0. |
- ConstantInstr* zero = |
- flow_graph()->GetConstant(Smi::Handle(I, Smi::New(0))); |
- *cursor = flow_graph()->AppendTo(*cursor, |
- new(I) CheckArrayBoundInstr( |
- new(I) Value(adjusted_length), |
- new(I) Value(zero), |
- call->deopt_id()), |
- call->env(), |
- FlowGraph::kEffect); |
- // Check 0 <= byte_index < adjusted_length. |
- *cursor = flow_graph()->AppendTo(*cursor, |
- new(I) CheckArrayBoundInstr( |
- new(I) Value(adjusted_length), |
- new(I) Value(byte_index), |
- call->deopt_id()), |
- call->env(), |
- FlowGraph::kEffect); |
- |
- if (RawObject::IsExternalTypedDataClassId(array_cid)) { |
- LoadUntaggedInstr* elements = |
- new(I) LoadUntaggedInstr(new(I) Value(*array), |
- ExternalTypedData::data_offset()); |
- *cursor = flow_graph()->AppendTo(*cursor, |
- elements, |
- NULL, |
- FlowGraph::kValue); |
- *array = elements; |
- } |
- return array_cid; |
-} |
- |
- |
-bool FlowGraphOptimizer::BuildByteArrayViewLoad(InstanceCallInstr* call, |
- intptr_t view_cid) { |
- const bool simd_view = (view_cid == kTypedDataFloat32x4ArrayCid) || |
- (view_cid == kTypedDataInt32x4ArrayCid); |
- const bool float_view = (view_cid == kTypedDataFloat32ArrayCid) || |
- (view_cid == kTypedDataFloat64ArrayCid); |
- if (float_view && !CanUnboxDouble()) { |
- return false; |
- } |
- if (simd_view && !ShouldInlineSimd()) { |
- return false; |
- } |
- return TryReplaceInstanceCallWithInline(call); |
-} |
- |
- |
-bool FlowGraphOptimizer::BuildByteArrayViewStore(InstanceCallInstr* call, |
- intptr_t view_cid) { |
- const bool simd_view = (view_cid == kTypedDataFloat32x4ArrayCid) || |
- (view_cid == kTypedDataInt32x4ArrayCid); |
- const bool float_view = (view_cid == kTypedDataFloat32ArrayCid) || |
- (view_cid == kTypedDataFloat64ArrayCid); |
- if (float_view && !CanUnboxDouble()) { |
- return false; |
- } |
- if (simd_view && !ShouldInlineSimd()) { |
- return false; |
- } |
- return TryReplaceInstanceCallWithInline(call); |
-} |
- |
- |
-// If type tests specified by 'ic_data' do not depend on type arguments, |
-// return mapping cid->result in 'results' (i : cid; i + 1: result). |
-// If all tests yield the same result, return it otherwise return Bool::null. |
-// If no mapping is possible, 'results' is empty. |
-// An instance-of test returning all same results can be converted to a class |
-// check. |
-RawBool* FlowGraphOptimizer::InstanceOfAsBool( |
- const ICData& ic_data, |
- const AbstractType& type, |
- ZoneGrowableArray<intptr_t>* results) const { |
- ASSERT(results->is_empty()); |
- ASSERT(ic_data.NumArgsTested() == 1); // Unary checks only. |
- if (!type.IsInstantiated() || type.IsMalformedOrMalbounded()) { |
- return Bool::null(); |
- } |
- const Class& type_class = Class::Handle(I, type.type_class()); |
- const intptr_t num_type_args = type_class.NumTypeArguments(); |
- if (num_type_args > 0) { |
- // Only raw types can be directly compared, thus disregarding type |
- // arguments. |
- const intptr_t num_type_params = type_class.NumTypeParameters(); |
- const intptr_t from_index = num_type_args - num_type_params; |
- const TypeArguments& type_arguments = |
- TypeArguments::Handle(I, type.arguments()); |
- const bool is_raw_type = type_arguments.IsNull() || |
- type_arguments.IsRaw(from_index, num_type_params); |
- if (!is_raw_type) { |
- // Unknown result. |
- return Bool::null(); |
- } |
- } |
- |
- const ClassTable& class_table = *isolate()->class_table(); |
- Bool& prev = Bool::Handle(I); |
- Class& cls = Class::Handle(I); |
- |
- bool results_differ = false; |
- for (int i = 0; i < ic_data.NumberOfChecks(); i++) { |
- cls = class_table.At(ic_data.GetReceiverClassIdAt(i)); |
- if (cls.NumTypeArguments() > 0) { |
- return Bool::null(); |
- } |
- const bool is_subtype = cls.IsSubtypeOf( |
- TypeArguments::Handle(I), |
- type_class, |
- TypeArguments::Handle(I), |
- NULL); |
- results->Add(cls.id()); |
- results->Add(is_subtype); |
- if (prev.IsNull()) { |
- prev = Bool::Get(is_subtype).raw(); |
- } else { |
- if (is_subtype != prev.value()) { |
- results_differ = true; |
- } |
- } |
- } |
- return results_differ ? Bool::null() : prev.raw(); |
-} |
- |
- |
-// Returns true if checking against this type is a direct class id comparison. |
-bool FlowGraphOptimizer::TypeCheckAsClassEquality(const AbstractType& type) { |
- ASSERT(type.IsFinalized() && !type.IsMalformedOrMalbounded()); |
- // Requires CHA. |
- if (!FLAG_use_cha) return false; |
- if (!type.IsInstantiated()) return false; |
- const Class& type_class = Class::Handle(type.type_class()); |
- // Signature classes have different type checking rules. |
- if (type_class.IsSignatureClass()) return false; |
- // Could be an interface check? |
- if (isolate()->cha()->IsImplemented(type_class)) return false; |
- // Check if there are subclasses. |
- if (isolate()->cha()->HasSubclasses(type_class)) return false; |
- const intptr_t num_type_args = type_class.NumTypeArguments(); |
- if (num_type_args > 0) { |
- // Only raw types can be directly compared, thus disregarding type |
- // arguments. |
- const intptr_t num_type_params = type_class.NumTypeParameters(); |
- const intptr_t from_index = num_type_args - num_type_params; |
- const TypeArguments& type_arguments = |
- TypeArguments::Handle(type.arguments()); |
- const bool is_raw_type = type_arguments.IsNull() || |
- type_arguments.IsRaw(from_index, num_type_params); |
- return is_raw_type; |
- } |
- return true; |
-} |
- |
- |
-static bool CidTestResultsContains(const ZoneGrowableArray<intptr_t>& results, |
- intptr_t test_cid) { |
- for (intptr_t i = 0; i < results.length(); i += 2) { |
- if (results[i] == test_cid) return true; |
- } |
- return false; |
-} |
- |
- |
-static void TryAddTest(ZoneGrowableArray<intptr_t>* results, |
- intptr_t test_cid, |
- bool result) { |
- if (!CidTestResultsContains(*results, test_cid)) { |
- results->Add(test_cid); |
- results->Add(result); |
- } |
-} |
- |
- |
-// Tries to add cid tests to 'results' so that no deoptimization is |
-// necessary. |
-// TODO(srdjan): Do also for other than 'int' type. |
-static bool TryExpandTestCidsResult(ZoneGrowableArray<intptr_t>* results, |
- const AbstractType& type) { |
- ASSERT(results->length() >= 2); // At least on eentry. |
- const ClassTable& class_table = *Isolate::Current()->class_table(); |
- if ((*results)[0] != kSmiCid) { |
- const Class& cls = Class::Handle(class_table.At(kSmiCid)); |
- const Class& type_class = Class::Handle(type.type_class()); |
- const bool smi_is_subtype = cls.IsSubtypeOf(TypeArguments::Handle(), |
- type_class, |
- TypeArguments::Handle(), |
- NULL); |
- results->Add((*results)[results->length() - 2]); |
- results->Add((*results)[results->length() - 2]); |
- for (intptr_t i = results->length() - 3; i > 1; --i) { |
- (*results)[i] = (*results)[i - 2]; |
- } |
- (*results)[0] = kSmiCid; |
- (*results)[1] = smi_is_subtype; |
- } |
- |
- ASSERT(type.IsInstantiated() && !type.IsMalformedOrMalbounded()); |
- ASSERT(results->length() >= 2); |
- if (type.IsIntType()) { |
- ASSERT((*results)[0] == kSmiCid); |
- TryAddTest(results, kMintCid, true); |
- TryAddTest(results, kBigintCid, true); |
- // Cannot deoptimize since all tests returning true have been added. |
- return false; |
- } |
- |
- return true; // May deoptimize since we have not identified all 'true' tests. |
-} |
- |
- |
-// TODO(srdjan): Use ICData to check if always true or false. |
-void FlowGraphOptimizer::ReplaceWithInstanceOf(InstanceCallInstr* call) { |
- ASSERT(Token::IsTypeTestOperator(call->token_kind())); |
- Definition* left = call->ArgumentAt(0); |
- Definition* instantiator = call->ArgumentAt(1); |
- Definition* type_args = call->ArgumentAt(2); |
- const AbstractType& type = |
- AbstractType::Cast(call->ArgumentAt(3)->AsConstant()->value()); |
- const bool negate = Bool::Cast( |
- call->ArgumentAt(4)->OriginalDefinition()->AsConstant()->value()).value(); |
- const ICData& unary_checks = |
- ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks()); |
- if (FLAG_warn_on_javascript_compatibility && |
- !unary_checks.IssuedJSWarning() && |
- (type.IsIntType() || type.IsDoubleType() || !type.IsInstantiated())) { |
- // No warning was reported yet for this type check, either because it has |
- // not been executed yet, or because no problematic combinations of instance |
- // type and test type have been encountered so far. A warning may still be |
- // reported, so do not replace the instance call. |
- return; |
- } |
- if (unary_checks.NumberOfChecks() <= FLAG_max_polymorphic_checks) { |
- ZoneGrowableArray<intptr_t>* results = |
- new(I) ZoneGrowableArray<intptr_t>(unary_checks.NumberOfChecks() * 2); |
- Bool& as_bool = |
- Bool::ZoneHandle(I, InstanceOfAsBool(unary_checks, type, results)); |
- if (as_bool.IsNull()) { |
- if (results->length() == unary_checks.NumberOfChecks() * 2) { |
- const bool can_deopt = TryExpandTestCidsResult(results, type); |
- TestCidsInstr* test_cids = new(I) TestCidsInstr( |
- call->token_pos(), |
- negate ? Token::kISNOT : Token::kIS, |
- new(I) Value(left), |
- *results, |
- can_deopt ? call->deopt_id() : Isolate::kNoDeoptId); |
- // Remove type. |
- ReplaceCall(call, test_cids); |
- return; |
- } |
- } else { |
- // TODO(srdjan): Use TestCidsInstr also for this case. |
- // One result only. |
- AddReceiverCheck(call); |
- if (negate) { |
- as_bool = Bool::Get(!as_bool.value()).raw(); |
- } |
- ConstantInstr* bool_const = flow_graph()->GetConstant(as_bool); |
- for (intptr_t i = 0; i < call->ArgumentCount(); ++i) { |
- PushArgumentInstr* push = call->PushArgumentAt(i); |
- push->ReplaceUsesWith(push->value()->definition()); |
- push->RemoveFromGraph(); |
- } |
- call->ReplaceUsesWith(bool_const); |
- ASSERT(current_iterator()->Current() == call); |
- current_iterator()->RemoveCurrentFromGraph(); |
- return; |
- } |
- } |
- |
- if (TypeCheckAsClassEquality(type)) { |
- LoadClassIdInstr* left_cid = new(I) LoadClassIdInstr(new(I) Value(left)); |
- InsertBefore(call, |
- left_cid, |
- NULL, |
- FlowGraph::kValue); |
- const intptr_t type_cid = Class::Handle(I, type.type_class()).id(); |
- ConstantInstr* cid = |
- flow_graph()->GetConstant(Smi::Handle(I, Smi::New(type_cid))); |
- |
- StrictCompareInstr* check_cid = |
- new(I) StrictCompareInstr( |
- call->token_pos(), |
- negate ? Token::kNE_STRICT : Token::kEQ_STRICT, |
- new(I) Value(left_cid), |
- new(I) Value(cid), |
- false); // No number check. |
- ReplaceCall(call, check_cid); |
- return; |
- } |
- |
- InstanceOfInstr* instance_of = |
- new(I) InstanceOfInstr(call->token_pos(), |
- new(I) Value(left), |
- new(I) Value(instantiator), |
- new(I) Value(type_args), |
- type, |
- negate, |
- call->deopt_id()); |
- ReplaceCall(call, instance_of); |
-} |
- |
- |
-// TODO(srdjan): Apply optimizations as in ReplaceWithInstanceOf (TestCids). |
-void FlowGraphOptimizer::ReplaceWithTypeCast(InstanceCallInstr* call) { |
- ASSERT(Token::IsTypeCastOperator(call->token_kind())); |
- Definition* left = call->ArgumentAt(0); |
- Definition* instantiator = call->ArgumentAt(1); |
- Definition* type_args = call->ArgumentAt(2); |
- const AbstractType& type = |
- AbstractType::Cast(call->ArgumentAt(3)->AsConstant()->value()); |
- ASSERT(!type.IsMalformedOrMalbounded()); |
- const ICData& unary_checks = |
- ICData::ZoneHandle(I, call->ic_data()->AsUnaryClassChecks()); |
- if (FLAG_warn_on_javascript_compatibility && |
- !unary_checks.IssuedJSWarning() && |
- (type.IsIntType() || type.IsDoubleType() || !type.IsInstantiated())) { |
- // No warning was reported yet for this type check, either because it has |
- // not been executed yet, or because no problematic combinations of instance |
- // type and test type have been encountered so far. A warning may still be |
- // reported, so do not replace the instance call. |
- return; |
- } |
- if (unary_checks.NumberOfChecks() <= FLAG_max_polymorphic_checks) { |
- ZoneGrowableArray<intptr_t>* results = |
- new(I) ZoneGrowableArray<intptr_t>(unary_checks.NumberOfChecks() * 2); |
- const Bool& as_bool = Bool::ZoneHandle(I, |
- InstanceOfAsBool(unary_checks, type, results)); |
- if (as_bool.raw() == Bool::True().raw()) { |
- AddReceiverCheck(call); |
- // Remove the original push arguments. |
- for (intptr_t i = 0; i < call->ArgumentCount(); ++i) { |
- PushArgumentInstr* push = call->PushArgumentAt(i); |
- push->ReplaceUsesWith(push->value()->definition()); |
- push->RemoveFromGraph(); |
- } |
- // Remove call, replace it with 'left'. |
- call->ReplaceUsesWith(left); |
- ASSERT(current_iterator()->Current() == call); |
- current_iterator()->RemoveCurrentFromGraph(); |
- return; |
- } |
- } |
- const String& dst_name = String::ZoneHandle(I, |
- Symbols::New(Exceptions::kCastErrorDstName)); |
- AssertAssignableInstr* assert_as = |
- new(I) AssertAssignableInstr(call->token_pos(), |
- new(I) Value(left), |
- new(I) Value(instantiator), |
- new(I) Value(type_args), |
- type, |
- dst_name, |
- call->deopt_id()); |
- ReplaceCall(call, assert_as); |
-} |
- |
- |
-// Tries to optimize instance call by replacing it with a faster instruction |
-// (e.g, binary op, field load, ..). |
-void FlowGraphOptimizer::VisitInstanceCall(InstanceCallInstr* instr) { |
- if (!instr->HasICData() || (instr->ic_data()->NumberOfUsedChecks() == 0)) { |
- return; |
- } |
- |
- const Token::Kind op_kind = instr->token_kind(); |
- // Type test is special as it always gets converted into inlined code. |
- if (Token::IsTypeTestOperator(op_kind)) { |
- ReplaceWithInstanceOf(instr); |
- return; |
- } |
- |
- if (Token::IsTypeCastOperator(op_kind)) { |
- ReplaceWithTypeCast(instr); |
- return; |
- } |
- |
- const ICData& unary_checks = |
- ICData::ZoneHandle(I, instr->ic_data()->AsUnaryClassChecks()); |
- |
- const intptr_t max_checks = (op_kind == Token::kEQ) |
- ? FLAG_max_equality_polymorphic_checks |
- : FLAG_max_polymorphic_checks; |
- if ((unary_checks.NumberOfChecks() > max_checks) && |
- InstanceCallNeedsClassCheck(instr, RawFunction::kRegularFunction)) { |
- // Too many checks, it will be megamorphic which needs unary checks. |
- instr->set_ic_data(&unary_checks); |
- return; |
- } |
- |
- if ((op_kind == Token::kASSIGN_INDEX) && TryReplaceWithStoreIndexed(instr)) { |
- return; |
- } |
- if ((op_kind == Token::kINDEX) && TryReplaceWithLoadIndexed(instr)) { |
- return; |
- } |
- |
- if (op_kind == Token::kEQ && TryReplaceWithEqualityOp(instr, op_kind)) { |
- return; |
- } |
- |
- if (Token::IsRelationalOperator(op_kind) && |
- TryReplaceWithRelationalOp(instr, op_kind)) { |
- return; |
- } |
- |
- if (Token::IsBinaryOperator(op_kind) && |
- TryReplaceWithBinaryOp(instr, op_kind)) { |
- return; |
- } |
- if (Token::IsUnaryOperator(op_kind) && |
- TryReplaceWithUnaryOp(instr, op_kind)) { |
- return; |
- } |
- if ((op_kind == Token::kGET) && TryInlineInstanceGetter(instr)) { |
- return; |
- } |
- if ((op_kind == Token::kSET) && |
- TryInlineInstanceSetter(instr, unary_checks)) { |
- return; |
- } |
- if (TryInlineInstanceMethod(instr)) { |
- return; |
- } |
- |
- bool has_one_target = unary_checks.HasOneTarget(); |
- |
- if (has_one_target) { |
- // Check if the single target is a polymorphic target, if it is, |
- // we don't have one target. |
- const Function& target = |
- Function::Handle(I, unary_checks.GetTargetAt(0)); |
- const bool polymorphic_target = MethodRecognizer::PolymorphicTarget(target); |
- has_one_target = !polymorphic_target; |
- } |
- |
- if (has_one_target) { |
- RawFunction::Kind function_kind = |
- Function::Handle(I, unary_checks.GetTargetAt(0)).kind(); |
- if (!InstanceCallNeedsClassCheck(instr, function_kind)) { |
- const bool call_with_checks = false; |
- PolymorphicInstanceCallInstr* call = |
- new(I) PolymorphicInstanceCallInstr(instr, unary_checks, |
- call_with_checks); |
- instr->ReplaceWith(call, current_iterator()); |
- return; |
- } |
- } |
- |
- if (unary_checks.NumberOfChecks() <= FLAG_max_polymorphic_checks) { |
- bool call_with_checks; |
- if (has_one_target) { |
- // Type propagation has not run yet, we cannot eliminate the check. |
- AddReceiverCheck(instr); |
- // Call can still deoptimize, do not detach environment from instr. |
- call_with_checks = false; |
- } else { |
- call_with_checks = true; |
- } |
- PolymorphicInstanceCallInstr* call = |
- new(I) PolymorphicInstanceCallInstr(instr, unary_checks, |
- call_with_checks); |
- instr->ReplaceWith(call, current_iterator()); |
- } |
-} |
- |
- |
-void FlowGraphOptimizer::VisitStaticCall(StaticCallInstr* call) { |
- if (!CanUnboxDouble()) { |
- return; |
- } |
- MethodRecognizer::Kind recognized_kind = |
- MethodRecognizer::RecognizeKind(call->function()); |
- MathUnaryInstr::MathUnaryKind unary_kind; |
- switch (recognized_kind) { |
- case MethodRecognizer::kMathSqrt: |
- unary_kind = MathUnaryInstr::kSqrt; |
- break; |
- case MethodRecognizer::kMathSin: |
- unary_kind = MathUnaryInstr::kSin; |
- break; |
- case MethodRecognizer::kMathCos: |
- unary_kind = MathUnaryInstr::kCos; |
- break; |
- default: |
- unary_kind = MathUnaryInstr::kIllegal; |
- break; |
- } |
- if (unary_kind != MathUnaryInstr::kIllegal) { |
- MathUnaryInstr* math_unary = |
- new(I) MathUnaryInstr(unary_kind, |
- new(I) Value(call->ArgumentAt(0)), |
- call->deopt_id()); |
- ReplaceCall(call, math_unary); |
- } else if ((recognized_kind == MethodRecognizer::kFloat32x4Zero) || |
- (recognized_kind == MethodRecognizer::kFloat32x4Splat) || |
- (recognized_kind == MethodRecognizer::kFloat32x4Constructor) || |
- (recognized_kind == MethodRecognizer::kFloat32x4FromFloat64x2)) { |
- TryInlineFloat32x4Constructor(call, recognized_kind); |
- } else if ((recognized_kind == MethodRecognizer::kFloat64x2Constructor) || |
- (recognized_kind == MethodRecognizer::kFloat64x2Zero) || |
- (recognized_kind == MethodRecognizer::kFloat64x2Splat) || |
- (recognized_kind == MethodRecognizer::kFloat64x2FromFloat32x4)) { |
- TryInlineFloat64x2Constructor(call, recognized_kind); |
- } else if ((recognized_kind == MethodRecognizer::kInt32x4BoolConstructor) || |
- (recognized_kind == MethodRecognizer::kInt32x4Constructor)) { |
- TryInlineInt32x4Constructor(call, recognized_kind); |
- } else if (recognized_kind == MethodRecognizer::kObjectConstructor) { |
- // Remove the original push arguments. |
- for (intptr_t i = 0; i < call->ArgumentCount(); ++i) { |
- PushArgumentInstr* push = call->PushArgumentAt(i); |
- push->ReplaceUsesWith(push->value()->definition()); |
- push->RemoveFromGraph(); |
- } |
- // Manually replace call with global null constant. ReplaceCall can't |
- // be used for definitions that are already in the graph. |
- call->ReplaceUsesWith(flow_graph_->constant_null()); |
- ASSERT(current_iterator()->Current() == call); |
- current_iterator()->RemoveCurrentFromGraph();; |
- } else if ((recognized_kind == MethodRecognizer::kMathMin) || |
- (recognized_kind == MethodRecognizer::kMathMax)) { |
- // We can handle only monomorphic min/max call sites with both arguments |
- // being either doubles or smis. |
- if (call->HasICData() && (call->ic_data()->NumberOfChecks() == 1)) { |
- const ICData& ic_data = *call->ic_data(); |
- intptr_t result_cid = kIllegalCid; |
- if (ICDataHasReceiverArgumentClassIds(ic_data, kDoubleCid, kDoubleCid)) { |
- result_cid = kDoubleCid; |
- } else if (ICDataHasReceiverArgumentClassIds(ic_data, kSmiCid, kSmiCid)) { |
- result_cid = kSmiCid; |
- } |
- if (result_cid != kIllegalCid) { |
- MathMinMaxInstr* min_max = new(I) MathMinMaxInstr( |
- recognized_kind, |
- new(I) Value(call->ArgumentAt(0)), |
- new(I) Value(call->ArgumentAt(1)), |
- call->deopt_id(), |
- result_cid); |
- const ICData& unary_checks = |
- ICData::ZoneHandle(I, ic_data.AsUnaryClassChecks()); |
- AddCheckClass(min_max->left()->definition(), |
- unary_checks, |
- call->deopt_id(), |
- call->env(), |
- call); |
- AddCheckClass(min_max->right()->definition(), |
- unary_checks, |
- call->deopt_id(), |
- call->env(), |
- call); |
- ReplaceCall(call, min_max); |
- } |
- } |
- } else if (recognized_kind == MethodRecognizer::kMathDoublePow) { |
- // We know that first argument is double, the second is num. |
- // InvokeMathCFunctionInstr requires unboxed doubles. UnboxDouble |
- // instructions contain type checks and conversions to double. |
- ZoneGrowableArray<Value*>* args = |
- new(I) ZoneGrowableArray<Value*>(call->ArgumentCount()); |
- for (intptr_t i = 0; i < call->ArgumentCount(); i++) { |
- args->Add(new(I) Value(call->ArgumentAt(i))); |
- } |
- InvokeMathCFunctionInstr* invoke = |
- new(I) InvokeMathCFunctionInstr(args, |
- call->deopt_id(), |
- recognized_kind, |
- call->token_pos()); |
- ReplaceCall(call, invoke); |
- } else if (recognized_kind == MethodRecognizer::kDoubleFromInteger) { |
- if (call->HasICData() && (call->ic_data()->NumberOfChecks() == 1)) { |
- const ICData& ic_data = *call->ic_data(); |
- if (CanUnboxDouble()) { |
- if (ArgIsAlways(kSmiCid, ic_data, 1)) { |
- Definition* arg = call->ArgumentAt(1); |
- AddCheckSmi(arg, call->deopt_id(), call->env(), call); |
- ReplaceCall(call, |
- new(I) SmiToDoubleInstr(new(I) Value(arg), |
- call->token_pos())); |
- } else if (ArgIsAlways(kMintCid, ic_data, 1) && |
- CanConvertUnboxedMintToDouble()) { |
- Definition* arg = call->ArgumentAt(1); |
- ReplaceCall(call, |
- new(I) MintToDoubleInstr(new(I) Value(arg), |
- call->deopt_id())); |
- } |
- } |
- } |
- } else if (call->function().IsFactory()) { |
- const Class& function_class = |
- Class::Handle(I, call->function().Owner()); |
- if ((function_class.library() == Library::CoreLibrary()) || |
- (function_class.library() == Library::TypedDataLibrary())) { |
- intptr_t cid = FactoryRecognizer::ResultCid(call->function()); |
- switch (cid) { |
- case kArrayCid: { |
- Value* type = new(I) Value(call->ArgumentAt(0)); |
- Value* num_elements = new(I) Value(call->ArgumentAt(1)); |
- if (num_elements->BindsToConstant() && |
- num_elements->BoundConstant().IsSmi()) { |
- intptr_t length = Smi::Cast(num_elements->BoundConstant()).Value(); |
- if (length >= 0 && length <= Array::kMaxElements) { |
- CreateArrayInstr* create_array = |
- new(I) CreateArrayInstr( |
- call->token_pos(), type, num_elements); |
- ReplaceCall(call, create_array); |
- } |
- } |
- } |
- default: |
- break; |
- } |
- } |
- } |
-} |
- |
- |
-void FlowGraphOptimizer::VisitStoreInstanceField( |
- StoreInstanceFieldInstr* instr) { |
- if (instr->IsUnboxedStore()) { |
- ASSERT(instr->is_initialization_); |
- // Determine if this field should be unboxed based on the usage of getter |
- // and setter functions: The heuristic requires that the setter has a |
- // usage count of at least 1/kGetterSetterRatio of the getter usage count. |
- // This is to avoid unboxing fields where the setter is never or rarely |
- // executed. |
- const Field& field = Field::ZoneHandle(I, instr->field().raw()); |
- const String& field_name = String::Handle(I, field.name()); |
- const Class& owner = Class::Handle(I, field.owner()); |
- const Function& getter = |
- Function::Handle(I, owner.LookupGetterFunction(field_name)); |
- const Function& setter = |
- Function::Handle(I, owner.LookupSetterFunction(field_name)); |
- bool result = !getter.IsNull() |
- && !setter.IsNull() |
- && (setter.usage_counter() > 0) |
- && (FLAG_getter_setter_ratio * setter.usage_counter() >= |
- getter.usage_counter()); |
- if (!result) { |
- if (FLAG_trace_optimization) { |
- OS::Print("Disabling unboxing of %s\n", field.ToCString()); |
- } |
- field.set_is_unboxing_candidate(false); |
- field.DeoptimizeDependentCode(); |
- } else { |
- FlowGraph::AddToGuardedFields(flow_graph_->guarded_fields(), &field); |
- } |
- } |
-} |
- |
- |
-void FlowGraphOptimizer::VisitAllocateContext(AllocateContextInstr* instr) { |
- // Replace generic allocation with a sequence of inlined allocation and |
- // explicit initalizing stores. |
- AllocateUninitializedContextInstr* replacement = |
- new AllocateUninitializedContextInstr(instr->token_pos(), |
- instr->num_context_variables()); |
- instr->ReplaceWith(replacement, current_iterator()); |
- |
- StoreInstanceFieldInstr* store = |
- new(I) StoreInstanceFieldInstr(Context::parent_offset(), |
- new Value(replacement), |
- new Value(flow_graph_->constant_null()), |
- kNoStoreBarrier, |
- instr->token_pos()); |
- store->set_is_initialization(true); // Won't be eliminated by DSE. |
- flow_graph_->InsertAfter(replacement, store, NULL, FlowGraph::kEffect); |
- Definition* cursor = store; |
- for (intptr_t i = 0; i < instr->num_context_variables(); ++i) { |
- store = |
- new(I) StoreInstanceFieldInstr(Context::variable_offset(i), |
- new Value(replacement), |
- new Value(flow_graph_->constant_null()), |
- kNoStoreBarrier, |
- instr->token_pos()); |
- store->set_is_initialization(true); // Won't be eliminated by DSE. |
- flow_graph_->InsertAfter(cursor, store, NULL, FlowGraph::kEffect); |
- cursor = store; |
- } |
-} |
- |
- |
-void FlowGraphOptimizer::VisitLoadCodeUnits(LoadCodeUnitsInstr* instr) { |
- // TODO(zerny): Use kUnboxedUint32 once it is fully supported/optimized. |
-#if defined(TARGET_ARCH_IA32) || defined(TARGET_ARCH_ARM) |
- if (!instr->can_pack_into_smi()) |
- instr->set_representation(kUnboxedMint); |
-#endif |
-} |
- |
- |
-bool FlowGraphOptimizer::TryInlineInstanceSetter(InstanceCallInstr* instr, |
- const ICData& unary_ic_data) { |
- ASSERT((unary_ic_data.NumberOfChecks() > 0) && |
- (unary_ic_data.NumArgsTested() == 1)); |
- if (FLAG_enable_type_checks) { |
- // Checked mode setters are inlined like normal methods by conventional |
- // inlining. |
- return false; |
- } |
- |
- ASSERT(instr->HasICData()); |
- if (unary_ic_data.NumberOfChecks() == 0) { |
- // No type feedback collected. |
- return false; |
- } |
- if (!unary_ic_data.HasOneTarget()) { |
- // Polymorphic sites are inlined like normal method calls by conventional |
- // inlining. |
- return false; |
- } |
- Function& target = Function::Handle(I); |
- intptr_t class_id; |
- unary_ic_data.GetOneClassCheckAt(0, &class_id, &target); |
- if (target.kind() != RawFunction::kImplicitSetter) { |
- // Non-implicit setter are inlined like normal method calls. |
- return false; |
- } |
- // Inline implicit instance setter. |
- const String& field_name = |
- String::Handle(I, Field::NameFromSetter(instr->function_name())); |
- const Field& field = |
- Field::ZoneHandle(I, GetField(class_id, field_name)); |
- ASSERT(!field.IsNull()); |
- |
- if (InstanceCallNeedsClassCheck(instr, RawFunction::kImplicitSetter)) { |
- AddReceiverCheck(instr); |
- } |
- StoreBarrierType needs_store_barrier = kEmitStoreBarrier; |
- if (ArgIsAlways(kSmiCid, *instr->ic_data(), 1)) { |
- InsertBefore(instr, |
- new(I) CheckSmiInstr( |
- new(I) Value(instr->ArgumentAt(1)), |
- instr->deopt_id(), |
- instr->token_pos()), |
- instr->env(), |
- FlowGraph::kEffect); |
- needs_store_barrier = kNoStoreBarrier; |
- } |
- |
- if (field.guarded_cid() != kDynamicCid) { |
- InsertBefore(instr, |
- new(I) GuardFieldClassInstr( |
- new(I) Value(instr->ArgumentAt(1)), |
- field, |
- instr->deopt_id()), |
- instr->env(), |
- FlowGraph::kEffect); |
- } |
- |
- if (field.needs_length_check()) { |
- InsertBefore(instr, |
- new(I) GuardFieldLengthInstr( |
- new(I) Value(instr->ArgumentAt(1)), |
- field, |
- instr->deopt_id()), |
- instr->env(), |
- FlowGraph::kEffect); |
- } |
- |
- // Field guard was detached. |
- StoreInstanceFieldInstr* store = new(I) StoreInstanceFieldInstr( |
- field, |
- new(I) Value(instr->ArgumentAt(0)), |
- new(I) Value(instr->ArgumentAt(1)), |
- needs_store_barrier, |
- instr->token_pos()); |
- |
- if (store->IsUnboxedStore()) { |
- FlowGraph::AddToGuardedFields(flow_graph_->guarded_fields(), &field); |
- } |
- |
- // Discard the environment from the original instruction because the store |
- // can't deoptimize. |
- instr->RemoveEnvironment(); |
- ReplaceCall(instr, store); |
- return true; |
-} |
- |
- |
-#if defined(TARGET_ARCH_ARM) || defined(TARGET_ARCH_IA32) |
-// Smi widening pass is only meaningful on platforms where Smi |
-// is smaller than 32bit. For now only support it on ARM and ia32. |
-static bool CanBeWidened(BinarySmiOpInstr* smi_op) { |
- return BinaryInt32OpInstr::IsSupported(smi_op->op_kind(), |
- smi_op->left(), |
- smi_op->right()); |
-} |
- |
- |
-static bool BenefitsFromWidening(BinarySmiOpInstr* smi_op) { |
- // TODO(vegorov): when shifts with non-constants shift count are supported |
- // add them here as we save untagging for the count. |
- switch (smi_op->op_kind()) { |
- case Token::kMUL: |
- case Token::kSHR: |
- // For kMUL we save untagging of the argument for kSHR |
- // we save tagging of the result. |
- return true; |
- |
- default: |
- return false; |
- } |
-} |
- |
- |
-void FlowGraphOptimizer::WidenSmiToInt32() { |
- GrowableArray<BinarySmiOpInstr*> candidates; |
- |
- // Step 1. Collect all instructions that potentially benefit from widening of |
- // their operands (or their result) into int32 range. |
- for (BlockIterator block_it = flow_graph_->reverse_postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- for (ForwardInstructionIterator instr_it(block_it.Current()); |
- !instr_it.Done(); |
- instr_it.Advance()) { |
- BinarySmiOpInstr* smi_op = instr_it.Current()->AsBinarySmiOp(); |
- if ((smi_op != NULL) && |
- BenefitsFromWidening(smi_op) && |
- CanBeWidened(smi_op)) { |
- candidates.Add(smi_op); |
- } |
- } |
- } |
- |
- if (candidates.is_empty()) { |
- return; |
- } |
- |
- // Step 2. For each block in the graph compute which loop it belongs to. |
- // We will use this information later during computation of the widening's |
- // gain: we are going to assume that only conversion occuring inside the |
- // same loop should be counted against the gain, all other conversions |
- // can be hoisted and thus cost nothing compared to the loop cost itself. |
- const ZoneGrowableArray<BlockEntryInstr*>& loop_headers = |
- flow_graph()->LoopHeaders(); |
- |
- GrowableArray<intptr_t> loops(flow_graph_->preorder().length()); |
- for (intptr_t i = 0; i < flow_graph_->preorder().length(); i++) { |
- loops.Add(-1); |
- } |
- |
- for (intptr_t loop_id = 0; loop_id < loop_headers.length(); ++loop_id) { |
- for (BitVector::Iterator loop_it(loop_headers[loop_id]->loop_info()); |
- !loop_it.Done(); |
- loop_it.Advance()) { |
- loops[loop_it.Current()] = loop_id; |
- } |
- } |
- |
- // Step 3. For each candidate transitively collect all other BinarySmiOpInstr |
- // and PhiInstr that depend on it and that it depends on and count amount of |
- // untagging operations that we save in assumption that this whole graph of |
- // values is using kUnboxedInt32 representation instead of kTagged. |
- // Convert those graphs that have positive gain to kUnboxedInt32. |
- |
- // BitVector containing SSA indexes of all processed definitions. Used to skip |
- // those candidates that belong to dependency graph of another candidate. |
- BitVector* processed = |
- new(I) BitVector(I, flow_graph_->current_ssa_temp_index()); |
- |
- // Worklist used to collect dependency graph. |
- DefinitionWorklist worklist(flow_graph_, candidates.length()); |
- for (intptr_t i = 0; i < candidates.length(); i++) { |
- BinarySmiOpInstr* op = candidates[i]; |
- if (op->WasEliminated() || processed->Contains(op->ssa_temp_index())) { |
- continue; |
- } |
- |
- if (FLAG_trace_smi_widening) { |
- OS::Print("analysing candidate: %s\n", op->ToCString()); |
- } |
- worklist.Clear(); |
- worklist.Add(op); |
- |
- // Collect dependency graph. Note: more items are added to worklist |
- // inside this loop. |
- intptr_t gain = 0; |
- for (intptr_t j = 0; j < worklist.definitions().length(); j++) { |
- Definition* defn = worklist.definitions()[j]; |
- |
- if (FLAG_trace_smi_widening) { |
- OS::Print("> %s\n", defn->ToCString()); |
- } |
- |
- if (defn->IsBinarySmiOp() && |
- BenefitsFromWidening(defn->AsBinarySmiOp())) { |
- gain++; |
- if (FLAG_trace_smi_widening) { |
- OS::Print("^ [%" Pd "] (o) %s\n", gain, defn->ToCString()); |
- } |
- } |
- |
- const intptr_t defn_loop = loops[defn->GetBlock()->preorder_number()]; |
- |
- // Process all inputs. |
- for (intptr_t k = 0; k < defn->InputCount(); k++) { |
- Definition* input = defn->InputAt(k)->definition(); |
- if (input->IsBinarySmiOp() && |
- CanBeWidened(input->AsBinarySmiOp())) { |
- worklist.Add(input); |
- } else if (input->IsPhi() && (input->Type()->ToCid() == kSmiCid)) { |
- worklist.Add(input); |
- } else if (input->IsBinaryMintOp()) { |
- // Mint operation produces untagged result. We avoid tagging. |
- gain++; |
- if (FLAG_trace_smi_widening) { |
- OS::Print("^ [%" Pd "] (i) %s\n", gain, input->ToCString()); |
- } |
- } else if (defn_loop == loops[input->GetBlock()->preorder_number()] && |
- (input->Type()->ToCid() == kSmiCid)) { |
- // Input comes from the same loop, is known to be smi and requires |
- // untagging. |
- // TODO(vegorov) this heuristic assumes that values that are not |
- // known to be smi have to be checked and this check can be |
- // coalesced with untagging. Start coalescing them. |
- gain--; |
- if (FLAG_trace_smi_widening) { |
- OS::Print("v [%" Pd "] (i) %s\n", gain, input->ToCString()); |
- } |
- } |
- } |
- |
- // Process all uses. |
- for (Value* use = defn->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- Instruction* instr = use->instruction(); |
- Definition* use_defn = instr->AsDefinition(); |
- if (use_defn == NULL) { |
- // We assume that tagging before returning or pushing argument costs |
- // very little compared to the cost of the return/call itself. |
- if (!instr->IsReturn() && !instr->IsPushArgument()) { |
- gain--; |
- if (FLAG_trace_smi_widening) { |
- OS::Print("v [%" Pd "] (u) %s\n", |
- gain, |
- use->instruction()->ToCString()); |
- } |
- } |
- continue; |
- } else if (use_defn->IsBinarySmiOp() && |
- CanBeWidened(use_defn->AsBinarySmiOp())) { |
- worklist.Add(use_defn); |
- } else if (use_defn->IsPhi() && |
- use_defn->AsPhi()->Type()->ToCid() == kSmiCid) { |
- worklist.Add(use_defn); |
- } else if (use_defn->IsBinaryMintOp()) { |
- // BinaryMintOp requires untagging of its inputs. |
- // Converting kUnboxedInt32 to kUnboxedMint is essentially zero cost |
- // sign extension operation. |
- gain++; |
- if (FLAG_trace_smi_widening) { |
- OS::Print("^ [%" Pd "] (u) %s\n", |
- gain, |
- use->instruction()->ToCString()); |
- } |
- } else if (defn_loop == loops[instr->GetBlock()->preorder_number()]) { |
- gain--; |
- if (FLAG_trace_smi_widening) { |
- OS::Print("v [%" Pd "] (u) %s\n", |
- gain, |
- use->instruction()->ToCString()); |
- } |
- } |
- } |
- } |
- |
- processed->AddAll(worklist.contains_vector()); |
- |
- if (FLAG_trace_smi_widening) { |
- OS::Print("~ %s gain %" Pd "\n", op->ToCString(), gain); |
- } |
- |
- if (gain > 0) { |
- // We have positive gain from widening. Convert all BinarySmiOpInstr into |
- // BinaryInt32OpInstr and set representation of all phis to kUnboxedInt32. |
- for (intptr_t j = 0; j < worklist.definitions().length(); j++) { |
- Definition* defn = worklist.definitions()[j]; |
- ASSERT(defn->IsPhi() || defn->IsBinarySmiOp()); |
- |
- if (defn->IsBinarySmiOp()) { |
- BinarySmiOpInstr* smi_op = defn->AsBinarySmiOp(); |
- BinaryInt32OpInstr* int32_op = new(I) BinaryInt32OpInstr( |
- smi_op->op_kind(), |
- smi_op->left()->CopyWithType(), |
- smi_op->right()->CopyWithType(), |
- smi_op->DeoptimizationTarget()); |
- |
- smi_op->ReplaceWith(int32_op, NULL); |
- } else if (defn->IsPhi()) { |
- defn->AsPhi()->set_representation(kUnboxedInt32); |
- ASSERT(defn->Type()->IsInt()); |
- } |
- } |
- } |
- } |
-} |
-#else |
-void FlowGraphOptimizer::WidenSmiToInt32() { |
- // TODO(vegorov) ideally on 64-bit platforms we would like to narrow smi |
- // operations to 32-bit where it saves tagging and untagging and allows |
- // to use shorted (and faster) instructions. But we currently don't |
- // save enough range information in the ICData to drive this decision. |
-} |
-#endif |
- |
-void FlowGraphOptimizer::InferIntRanges() { |
- RangeAnalysis range_analysis(flow_graph_); |
- range_analysis.Analyze(); |
-} |
- |
- |
-void TryCatchAnalyzer::Optimize(FlowGraph* flow_graph) { |
- // For every catch-block: Iterate over all call instructions inside the |
- // corresponding try-block and figure out for each environment value if it |
- // is the same constant at all calls. If yes, replace the initial definition |
- // at the catch-entry with this constant. |
- const GrowableArray<CatchBlockEntryInstr*>& catch_entries = |
- flow_graph->graph_entry()->catch_entries(); |
- intptr_t base = kFirstLocalSlotFromFp + flow_graph->num_non_copied_params(); |
- for (intptr_t catch_idx = 0; |
- catch_idx < catch_entries.length(); |
- ++catch_idx) { |
- CatchBlockEntryInstr* catch_entry = catch_entries[catch_idx]; |
- |
- // Initialize cdefs with the original initial definitions (ParameterInstr). |
- // The following representation is used: |
- // ParameterInstr => unknown |
- // ConstantInstr => known constant |
- // NULL => non-constant |
- GrowableArray<Definition*>* idefs = catch_entry->initial_definitions(); |
- GrowableArray<Definition*> cdefs(idefs->length()); |
- cdefs.AddArray(*idefs); |
- |
- // exception_var and stacktrace_var are never constant. |
- intptr_t ex_idx = base - catch_entry->exception_var().index(); |
- intptr_t st_idx = base - catch_entry->stacktrace_var().index(); |
- cdefs[ex_idx] = cdefs[st_idx] = NULL; |
- |
- for (BlockIterator block_it = flow_graph->reverse_postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- if (block->try_index() == catch_entry->catch_try_index()) { |
- for (ForwardInstructionIterator instr_it(block); |
- !instr_it.Done(); |
- instr_it.Advance()) { |
- Instruction* current = instr_it.Current(); |
- if (current->MayThrow()) { |
- Environment* env = current->env()->Outermost(); |
- ASSERT(env != NULL); |
- for (intptr_t env_idx = 0; env_idx < cdefs.length(); ++env_idx) { |
- if (cdefs[env_idx] != NULL && |
- env->ValueAt(env_idx)->BindsToConstant()) { |
- cdefs[env_idx] = env->ValueAt(env_idx)->definition(); |
- } |
- if (cdefs[env_idx] != env->ValueAt(env_idx)->definition()) { |
- cdefs[env_idx] = NULL; |
- } |
- } |
- } |
- } |
- } |
- } |
- for (intptr_t j = 0; j < idefs->length(); ++j) { |
- if (cdefs[j] != NULL && cdefs[j]->IsConstant()) { |
- // TODO(fschneider): Use constants from the constant pool. |
- Definition* old = (*idefs)[j]; |
- ConstantInstr* orig = cdefs[j]->AsConstant(); |
- ConstantInstr* copy = |
- new(flow_graph->isolate()) ConstantInstr(orig->value()); |
- copy->set_ssa_temp_index(flow_graph->alloc_ssa_temp_index()); |
- old->ReplaceUsesWith(copy); |
- (*idefs)[j] = copy; |
- } |
- } |
- } |
-} |
- |
- |
-LICM::LICM(FlowGraph* flow_graph) : flow_graph_(flow_graph) { |
- ASSERT(flow_graph->is_licm_allowed()); |
-} |
- |
- |
-void LICM::Hoist(ForwardInstructionIterator* it, |
- BlockEntryInstr* pre_header, |
- Instruction* current) { |
- if (current->IsCheckClass()) { |
- current->AsCheckClass()->set_licm_hoisted(true); |
- } else if (current->IsCheckSmi()) { |
- current->AsCheckSmi()->set_licm_hoisted(true); |
- } else if (current->IsCheckEitherNonSmi()) { |
- current->AsCheckEitherNonSmi()->set_licm_hoisted(true); |
- } else if (current->IsCheckArrayBound()) { |
- current->AsCheckArrayBound()->set_licm_hoisted(true); |
- } |
- if (FLAG_trace_optimization) { |
- OS::Print("Hoisting instruction %s:%" Pd " from B%" Pd " to B%" Pd "\n", |
- current->DebugName(), |
- current->GetDeoptId(), |
- current->GetBlock()->block_id(), |
- pre_header->block_id()); |
- } |
- // Move the instruction out of the loop. |
- current->RemoveEnvironment(); |
- if (it != NULL) { |
- it->RemoveCurrentFromGraph(); |
- } else { |
- current->RemoveFromGraph(); |
- } |
- GotoInstr* last = pre_header->last_instruction()->AsGoto(); |
- // Using kind kEffect will not assign a fresh ssa temporary index. |
- flow_graph()->InsertBefore(last, current, last->env(), FlowGraph::kEffect); |
- current->CopyDeoptIdFrom(*last); |
-} |
- |
- |
-void LICM::TrySpecializeSmiPhi(PhiInstr* phi, |
- BlockEntryInstr* header, |
- BlockEntryInstr* pre_header) { |
- if (phi->Type()->ToCid() == kSmiCid) { |
- return; |
- } |
- |
- // Check if there is only a single kDynamicCid input to the phi that |
- // comes from the pre-header. |
- const intptr_t kNotFound = -1; |
- intptr_t non_smi_input = kNotFound; |
- for (intptr_t i = 0; i < phi->InputCount(); ++i) { |
- Value* input = phi->InputAt(i); |
- if (input->Type()->ToCid() != kSmiCid) { |
- if ((non_smi_input != kNotFound) || |
- (input->Type()->ToCid() != kDynamicCid)) { |
- // There are multiple kDynamicCid inputs or there is an input that is |
- // known to be non-smi. |
- return; |
- } else { |
- non_smi_input = i; |
- } |
- } |
- } |
- |
- if ((non_smi_input == kNotFound) || |
- (phi->block()->PredecessorAt(non_smi_input) != pre_header)) { |
- return; |
- } |
- |
- CheckSmiInstr* check = NULL; |
- for (Value* use = phi->input_use_list(); |
- (use != NULL) && (check == NULL); |
- use = use->next_use()) { |
- check = use->instruction()->AsCheckSmi(); |
- } |
- |
- if (check == NULL) { |
- return; |
- } |
- |
- // Host CheckSmi instruction and make this phi smi one. |
- Hoist(NULL, pre_header, check); |
- |
- // Replace value we are checking with phi's input. |
- check->value()->BindTo(phi->InputAt(non_smi_input)->definition()); |
- |
- phi->UpdateType(CompileType::FromCid(kSmiCid)); |
-} |
- |
- |
-// Load instructions handled by load elimination. |
-static bool IsLoadEliminationCandidate(Instruction* instr) { |
- return instr->IsLoadField() |
- || instr->IsLoadIndexed() |
- || instr->IsLoadStaticField(); |
-} |
- |
- |
-static bool IsLoopInvariantLoad(ZoneGrowableArray<BitVector*>* sets, |
- intptr_t loop_header_index, |
- Instruction* instr) { |
- return IsLoadEliminationCandidate(instr) && |
- (sets != NULL) && |
- instr->HasPlaceId() && |
- ((*sets)[loop_header_index] != NULL) && |
- (*sets)[loop_header_index]->Contains(instr->place_id()); |
-} |
- |
- |
-void LICM::OptimisticallySpecializeSmiPhis() { |
- if (!flow_graph()->parsed_function()->function(). |
- allows_hoisting_check_class()) { |
- // Do not hoist any. |
- return; |
- } |
- |
- const ZoneGrowableArray<BlockEntryInstr*>& loop_headers = |
- flow_graph()->LoopHeaders(); |
- |
- for (intptr_t i = 0; i < loop_headers.length(); ++i) { |
- JoinEntryInstr* header = loop_headers[i]->AsJoinEntry(); |
- // Skip loop that don't have a pre-header block. |
- BlockEntryInstr* pre_header = header->ImmediateDominator(); |
- if (pre_header == NULL) continue; |
- |
- for (PhiIterator it(header); !it.Done(); it.Advance()) { |
- TrySpecializeSmiPhi(it.Current(), header, pre_header); |
- } |
- } |
-} |
- |
- |
-void LICM::Optimize() { |
- if (!flow_graph()->parsed_function()->function(). |
- allows_hoisting_check_class()) { |
- // Do not hoist any. |
- return; |
- } |
- |
- const ZoneGrowableArray<BlockEntryInstr*>& loop_headers = |
- flow_graph()->LoopHeaders(); |
- |
- ZoneGrowableArray<BitVector*>* loop_invariant_loads = |
- flow_graph()->loop_invariant_loads(); |
- |
- BlockEffects* block_effects = flow_graph()->block_effects(); |
- |
- for (intptr_t i = 0; i < loop_headers.length(); ++i) { |
- BlockEntryInstr* header = loop_headers[i]; |
- // Skip loop that don't have a pre-header block. |
- BlockEntryInstr* pre_header = header->ImmediateDominator(); |
- if (pre_header == NULL) continue; |
- |
- for (BitVector::Iterator loop_it(header->loop_info()); |
- !loop_it.Done(); |
- loop_it.Advance()) { |
- BlockEntryInstr* block = flow_graph()->preorder()[loop_it.Current()]; |
- for (ForwardInstructionIterator it(block); |
- !it.Done(); |
- it.Advance()) { |
- Instruction* current = it.Current(); |
- if ((current->AllowsCSE() && |
- block_effects->CanBeMovedTo(current, pre_header)) || |
- IsLoopInvariantLoad(loop_invariant_loads, i, current)) { |
- bool inputs_loop_invariant = true; |
- for (int i = 0; i < current->InputCount(); ++i) { |
- Definition* input_def = current->InputAt(i)->definition(); |
- if (!input_def->GetBlock()->Dominates(pre_header)) { |
- inputs_loop_invariant = false; |
- break; |
- } |
- } |
- if (inputs_loop_invariant && |
- !current->IsAssertAssignable() && |
- !current->IsAssertBoolean()) { |
- // TODO(fschneider): Enable hoisting of Assert-instructions |
- // if it safe to do. |
- Hoist(&it, pre_header, current); |
- } |
- } |
- } |
- } |
- } |
-} |
- |
- |
-// Place describes an abstract location (e.g. field) that IR can load |
-// from or store to. |
-// |
-// Places are also used to describe wild-card locations also known as aliases, |
-// that essentially represent sets of places that alias each other. Places A |
-// and B are said to alias each other if store into A can affect load from B. |
-// |
-// We distinguish the following aliases: |
-// |
-// - for fields |
-// - *.f, *.@offs - field inside some object; |
-// - X.f, X.@offs - field inside an allocated object X; |
-// - for indexed accesses |
-// - *[*] - non-constant index inside some object; |
-// - *[C] - constant index inside some object; |
-// - X[*] - non-constant index inside an allocated object X; |
-// - X[C] - constant index inside an allocated object X. |
-// |
-// Separating allocations from other objects improves precision of the |
-// load forwarding pass because of the following two properties: |
-// |
-// - if X can be proven to have no aliases itself (i.e. there is no other SSA |
-// variable that points to X) then no place inside X can be aliased with any |
-// wildcard dependent place (*.f, *.@offs, *[*], *[C]); |
-// - given allocations X and Y no place inside X can be aliased with any place |
-// inside Y even if any of them or both escape. |
-// |
-// It important to realize that single place can belong to multiple aliases. |
-// For example place X.f with aliased allocation X belongs both to X.f and *.f |
-// aliases. Likewise X[C] with non-aliased allocation X belongs to X[C] and X[*] |
-// aliases. |
-// |
-class Place : public ValueObject { |
- public: |
- enum Kind { |
- kNone, |
- |
- // Field location. For instance fields is represented as a pair of a Field |
- // object and an instance (SSA definition) that is being accessed. |
- // For static fields instance is NULL. |
- kField, |
- |
- // VMField location. Represented as a pair of an instance (SSA definition) |
- // being accessed and offset to the field. |
- kVMField, |
- |
- // Indexed location with a non-constant index. |
- kIndexed, |
- |
- // Indexed location with a constant index. |
- kConstantIndexed, |
- }; |
- |
- Place(const Place& other) |
- : ValueObject(), |
- kind_(other.kind_), |
- representation_(other.representation_), |
- instance_(other.instance_), |
- raw_selector_(other.raw_selector_), |
- id_(other.id_) { |
- } |
- |
- // Construct a place from instruction if instruction accesses any place. |
- // Otherwise constructs kNone place. |
- Place(Instruction* instr, bool* is_load, bool* is_store) |
- : kind_(kNone), |
- representation_(kNoRepresentation), |
- instance_(NULL), |
- raw_selector_(0), |
- id_(0) { |
- switch (instr->tag()) { |
- case Instruction::kLoadField: { |
- LoadFieldInstr* load_field = instr->AsLoadField(); |
- representation_ = load_field->representation(); |
- instance_ = load_field->instance()->definition()->OriginalDefinition(); |
- if (load_field->field() != NULL) { |
- kind_ = kField; |
- field_ = load_field->field(); |
- } else { |
- kind_ = kVMField; |
- offset_in_bytes_ = load_field->offset_in_bytes(); |
- } |
- *is_load = true; |
- break; |
- } |
- |
- case Instruction::kStoreInstanceField: { |
- StoreInstanceFieldInstr* store = |
- instr->AsStoreInstanceField(); |
- representation_ = store->RequiredInputRepresentation( |
- StoreInstanceFieldInstr::kValuePos); |
- instance_ = store->instance()->definition()->OriginalDefinition(); |
- if (!store->field().IsNull()) { |
- kind_ = kField; |
- field_ = &store->field(); |
- } else { |
- kind_ = kVMField; |
- offset_in_bytes_ = store->offset_in_bytes(); |
- } |
- *is_store = true; |
- break; |
- } |
- |
- case Instruction::kLoadStaticField: |
- kind_ = kField; |
- representation_ = instr->AsLoadStaticField()->representation(); |
- field_ = &instr->AsLoadStaticField()->StaticField(); |
- *is_load = true; |
- break; |
- |
- case Instruction::kStoreStaticField: |
- kind_ = kField; |
- representation_ = instr->AsStoreStaticField()-> |
- RequiredInputRepresentation(StoreStaticFieldInstr::kValuePos); |
- field_ = &instr->AsStoreStaticField()->field(); |
- *is_store = true; |
- break; |
- |
- case Instruction::kLoadIndexed: { |
- LoadIndexedInstr* load_indexed = instr->AsLoadIndexed(); |
- representation_ = load_indexed->representation(); |
- instance_ = load_indexed->array()->definition()->OriginalDefinition(); |
- SetIndex(load_indexed->index()->definition()); |
- *is_load = true; |
- break; |
- } |
- |
- case Instruction::kStoreIndexed: { |
- StoreIndexedInstr* store_indexed = instr->AsStoreIndexed(); |
- representation_ = store_indexed-> |
- RequiredInputRepresentation(StoreIndexedInstr::kValuePos); |
- instance_ = store_indexed->array()->definition()->OriginalDefinition(); |
- SetIndex(store_indexed->index()->definition()); |
- *is_store = true; |
- break; |
- } |
- |
- default: |
- break; |
- } |
- } |
- |
- // Create object representing *[*] alias. |
- static Place* CreateAnyInstanceAnyIndexAlias(Isolate* isolate, |
- intptr_t id) { |
- return Wrap(isolate, Place(kIndexed, NULL, 0), id); |
- } |
- |
- // Return least generic alias for this place. Given that aliases are |
- // essentially sets of places we define least generic alias as a smallest |
- // alias that contains this place. |
- // |
- // We obtain such alias by a simple transformation: |
- // |
- // - for places that depend on an instance X.f, X.@offs, X[i], X[C] |
- // we drop X if X is not an allocation because in this case X does not |
- // posess an identity obtaining aliases *.f, *.@offs, *[i] and *[C] |
- // respectively; |
- // - for non-constant indexed places X[i] we drop information about the |
- // index obtaining alias X[*]. |
- // |
- Place ToAlias() const { |
- return Place( |
- kind_, |
- (DependsOnInstance() && IsAllocation(instance())) ? instance() : NULL, |
- (kind() == kIndexed) ? 0 : raw_selector_); |
- } |
- |
- bool DependsOnInstance() const { |
- switch (kind()) { |
- case kField: |
- case kVMField: |
- case kIndexed: |
- case kConstantIndexed: |
- return true; |
- |
- case kNone: |
- return false; |
- } |
- |
- UNREACHABLE(); |
- return false; |
- } |
- |
- // Given instance dependent alias X.f, X.@offs, X[C], X[*] return |
- // wild-card dependent alias *.f, *.@offs, *[C] or *[*] respectively. |
- Place CopyWithoutInstance() const { |
- ASSERT(DependsOnInstance()); |
- return Place(kind_, NULL, raw_selector_); |
- } |
- |
- // Given alias X[C] or *[C] return X[*] and *[*] respectively. |
- Place CopyWithoutIndex() const { |
- ASSERT(kind_ == kConstantIndexed); |
- return Place(kIndexed, instance_, 0); |
- } |
- |
- intptr_t id() const { return id_; } |
- |
- Kind kind() const { return kind_; } |
- |
- Representation representation() const { return representation_; } |
- |
- Definition* instance() const { |
- ASSERT(DependsOnInstance()); |
- return instance_; |
- } |
- |
- void set_instance(Definition* def) { |
- ASSERT(DependsOnInstance()); |
- instance_ = def->OriginalDefinition(); |
- } |
- |
- const Field& field() const { |
- ASSERT(kind_ == kField); |
- return *field_; |
- } |
- |
- intptr_t offset_in_bytes() const { |
- ASSERT(kind_ == kVMField); |
- return offset_in_bytes_; |
- } |
- |
- Definition* index() const { |
- ASSERT(kind_ == kIndexed); |
- return index_; |
- } |
- |
- intptr_t index_constant() const { |
- ASSERT(kind_ == kConstantIndexed); |
- return index_constant_; |
- } |
- |
- static const char* DefinitionName(Definition* def) { |
- if (def == NULL) { |
- return "*"; |
- } else { |
- return Isolate::Current()->current_zone()->PrintToString( |
- "v%" Pd, def->ssa_temp_index()); |
- } |
- } |
- |
- const char* ToCString() const { |
- switch (kind_) { |
- case kNone: |
- return "<none>"; |
- |
- case kField: { |
- const char* field_name = String::Handle(field().name()).ToCString(); |
- if (field().is_static()) { |
- return Isolate::Current()->current_zone()->PrintToString( |
- "<%s>", field_name); |
- } else { |
- return Isolate::Current()->current_zone()->PrintToString( |
- "<%s.%s>", DefinitionName(instance()), field_name); |
- } |
- } |
- |
- case kVMField: |
- return Isolate::Current()->current_zone()->PrintToString( |
- "<%s.@%" Pd ">", |
- DefinitionName(instance()), |
- offset_in_bytes()); |
- |
- case kIndexed: |
- return Isolate::Current()->current_zone()->PrintToString( |
- "<%s[%s]>", |
- DefinitionName(instance()), |
- DefinitionName(index())); |
- |
- case kConstantIndexed: |
- return Isolate::Current()->current_zone()->PrintToString( |
- "<%s[%" Pd "]>", |
- DefinitionName(instance()), |
- index_constant()); |
- } |
- UNREACHABLE(); |
- return "<?>"; |
- } |
- |
- bool IsFinalField() const { |
- return (kind() == kField) && field().is_final(); |
- } |
- |
- intptr_t Hashcode() const { |
- return (kind_ * 63 + reinterpret_cast<intptr_t>(instance_)) * 31 + |
- representation_ * 15 + FieldHashcode(); |
- } |
- |
- bool Equals(const Place* other) const { |
- return (kind_ == other->kind_) && |
- (representation_ == other->representation_) && |
- (instance_ == other->instance_) && |
- SameField(other); |
- } |
- |
- // Create a zone allocated copy of this place and assign given id to it. |
- static Place* Wrap(Isolate* isolate, const Place& place, intptr_t id); |
- |
- static bool IsAllocation(Definition* defn) { |
- return (defn != NULL) && |
- (defn->IsAllocateObject() || |
- defn->IsCreateArray() || |
- defn->IsAllocateUninitializedContext() || |
- (defn->IsStaticCall() && |
- defn->AsStaticCall()->IsRecognizedFactory())); |
- } |
- |
- private: |
- Place(Kind kind, Definition* instance, intptr_t selector) |
- : kind_(kind), |
- representation_(kNoRepresentation), |
- instance_(instance), |
- raw_selector_(selector), |
- id_(0) { |
- } |
- |
- bool SameField(const Place* other) const { |
- return (kind_ == kField) ? (field().raw() == other->field().raw()) |
- : (offset_in_bytes_ == other->offset_in_bytes_); |
- } |
- |
- intptr_t FieldHashcode() const { |
- return (kind_ == kField) ? reinterpret_cast<intptr_t>(field().raw()) |
- : offset_in_bytes_; |
- } |
- |
- void SetIndex(Definition* index) { |
- ConstantInstr* index_constant = index->AsConstant(); |
- if ((index_constant != NULL) && index_constant->value().IsSmi()) { |
- kind_ = kConstantIndexed; |
- index_constant_ = Smi::Cast(index_constant->value()).Value(); |
- } else { |
- kind_ = kIndexed; |
- index_ = index; |
- } |
- } |
- |
- Kind kind_; |
- Representation representation_; |
- Definition* instance_; |
- union { |
- intptr_t raw_selector_; |
- const Field* field_; |
- intptr_t offset_in_bytes_; |
- intptr_t index_constant_; |
- Definition* index_; |
- }; |
- |
- intptr_t id_; |
-}; |
- |
- |
-class ZonePlace : public ZoneAllocated { |
- public: |
- explicit ZonePlace(const Place& place) : place_(place) { } |
- |
- Place* place() { return &place_; } |
- |
- private: |
- Place place_; |
-}; |
- |
- |
-Place* Place::Wrap(Isolate* isolate, const Place& place, intptr_t id) { |
- Place* wrapped = (new(isolate) ZonePlace(place))->place(); |
- wrapped->id_ = id; |
- return wrapped; |
-} |
- |
- |
-// Correspondence between places connected through outgoing phi moves on the |
-// edge that targets join. |
-class PhiPlaceMoves : public ZoneAllocated { |
- public: |
- // Record a move from the place with id |from| to the place with id |to| at |
- // the given block. |
- void CreateOutgoingMove(Isolate* isolate, |
- BlockEntryInstr* block, intptr_t from, intptr_t to) { |
- const intptr_t block_num = block->preorder_number(); |
- while (moves_.length() <= block_num) { |
- moves_.Add(NULL); |
- } |
- |
- if (moves_[block_num] == NULL) { |
- moves_[block_num] = new(isolate) ZoneGrowableArray<Move>(5); |
- } |
- |
- moves_[block_num]->Add(Move(from, to)); |
- } |
- |
- class Move { |
- public: |
- Move(intptr_t from, intptr_t to) : from_(from), to_(to) { } |
- |
- intptr_t from() const { return from_; } |
- intptr_t to() const { return to_; } |
- |
- private: |
- intptr_t from_; |
- intptr_t to_; |
- }; |
- |
- typedef const ZoneGrowableArray<Move>* MovesList; |
- |
- MovesList GetOutgoingMoves(BlockEntryInstr* block) const { |
- const intptr_t block_num = block->preorder_number(); |
- return (block_num < moves_.length()) ? |
- moves_[block_num] : NULL; |
- } |
- |
- private: |
- GrowableArray<ZoneGrowableArray<Move>* > moves_; |
-}; |
- |
- |
-// A map from aliases to a set of places sharing the alias. Additionally |
-// carries a set of places that can be aliased by side-effects, essentially |
-// those that are affected by calls. |
-class AliasedSet : public ZoneAllocated { |
- public: |
- AliasedSet(Isolate* isolate, |
- DirectChainedHashMap<PointerKeyValueTrait<Place> >* places_map, |
- ZoneGrowableArray<Place*>* places, |
- PhiPlaceMoves* phi_moves) |
- : isolate_(isolate), |
- places_map_(places_map), |
- places_(*places), |
- phi_moves_(phi_moves), |
- aliases_(5), |
- aliases_map_(), |
- representatives_(), |
- killed_(), |
- aliased_by_effects_(new(isolate) BitVector(isolate, places->length())) { |
- InsertAlias(Place::CreateAnyInstanceAnyIndexAlias(isolate_, |
- kAnyInstanceAnyIndexAlias)); |
- for (intptr_t i = 0; i < places_.length(); i++) { |
- AddRepresentative(places_[i]); |
- } |
- ComputeKillSets(); |
- } |
- |
- intptr_t LookupAliasId(const Place& alias) { |
- const Place* result = aliases_map_.Lookup(&alias); |
- return (result != NULL) ? result->id() : static_cast<intptr_t>(kNoAlias); |
- } |
- |
- BitVector* GetKilledSet(intptr_t alias) { |
- return (alias < killed_.length()) ? killed_[alias] : NULL; |
- } |
- |
- intptr_t max_place_id() const { return places().length(); } |
- bool IsEmpty() const { return max_place_id() == 0; } |
- |
- BitVector* aliased_by_effects() const { return aliased_by_effects_; } |
- |
- const ZoneGrowableArray<Place*>& places() const { |
- return places_; |
- } |
- |
- Place* LookupCanonical(Place* place) const { |
- return places_map_->Lookup(place); |
- } |
- |
- void PrintSet(BitVector* set) { |
- bool comma = false; |
- for (BitVector::Iterator it(set); |
- !it.Done(); |
- it.Advance()) { |
- if (comma) { |
- OS::Print(", "); |
- } |
- OS::Print("%s", places_[it.Current()]->ToCString()); |
- comma = true; |
- } |
- } |
- |
- const PhiPlaceMoves* phi_moves() const { return phi_moves_; } |
- |
- void RollbackAliasedIdentites() { |
- for (intptr_t i = 0; i < identity_rollback_.length(); ++i) { |
- identity_rollback_[i]->SetIdentity(AliasIdentity::Unknown()); |
- } |
- } |
- |
- // Returns false if the result of an allocation instruction can't be aliased |
- // by another SSA variable and true otherwise. |
- bool CanBeAliased(Definition* alloc) { |
- if (!Place::IsAllocation(alloc)) { |
- return true; |
- } |
- |
- if (alloc->Identity().IsUnknown()) { |
- ComputeAliasing(alloc); |
- } |
- |
- return !alloc->Identity().IsNotAliased(); |
- } |
- |
- enum { |
- kNoAlias = 0 |
- }; |
- |
- private: |
- enum { |
- // Artificial alias that is used to collect all representatives of the |
- // *[C], X[C] aliases for arbitrary C. |
- kAnyConstantIndexedAlias = 1, |
- |
- // Artificial alias that is used to collect all representatives of |
- // *[C] alias for arbitrary C. |
- kUnknownInstanceConstantIndexedAlias = 2, |
- |
- // Artificial alias that is used to collect all representatives of |
- // X[*] alias for all X. |
- kAnyAllocationIndexedAlias = 3, |
- |
- // *[*] alias. |
- kAnyInstanceAnyIndexAlias = 4 |
- }; |
- |
- // Compute least generic alias for the place and assign alias id to it. |
- void AddRepresentative(Place* place) { |
- if (!place->IsFinalField()) { |
- const Place* alias = CanonicalizeAlias(place->ToAlias()); |
- EnsureSet(&representatives_, alias->id())->Add(place->id()); |
- |
- // Update cumulative representative sets that are used during |
- // killed sets computation. |
- if (alias->kind() == Place::kConstantIndexed) { |
- if (CanBeAliased(alias->instance())) { |
- EnsureSet(&representatives_, kAnyConstantIndexedAlias)-> |
- Add(place->id()); |
- } |
- |
- if (alias->instance() == NULL) { |
- EnsureSet(&representatives_, kUnknownInstanceConstantIndexedAlias)-> |
- Add(place->id()); |
- } |
- } else if ((alias->kind() == Place::kIndexed) && |
- CanBeAliased(place->instance())) { |
- EnsureSet(&representatives_, kAnyAllocationIndexedAlias)-> |
- Add(place->id()); |
- } |
- |
- if (!IsIndependentFromEffects(place)) { |
- aliased_by_effects_->Add(place->id()); |
- } |
- } |
- } |
- |
- void ComputeKillSets() { |
- for (intptr_t i = 0; i < aliases_.length(); ++i) { |
- const Place* alias = aliases_[i]; |
- // Add all representatives to the kill set. |
- AddAllRepresentatives(alias->id(), alias->id()); |
- ComputeKillSet(alias); |
- } |
- |
- if (FLAG_trace_load_optimization) { |
- OS::Print("Aliases KILL sets:\n"); |
- for (intptr_t i = 0; i < aliases_.length(); ++i) { |
- const Place* alias = aliases_[i]; |
- BitVector* kill = GetKilledSet(alias->id()); |
- |
- OS::Print("%s: ", alias->ToCString()); |
- if (kill != NULL) { |
- PrintSet(kill); |
- } |
- OS::Print("\n"); |
- } |
- } |
- } |
- |
- void InsertAlias(const Place* alias) { |
- aliases_map_.Insert(alias); |
- aliases_.Add(alias); |
- } |
- |
- const Place* CanonicalizeAlias(const Place& alias) { |
- const Place* canonical = aliases_map_.Lookup(&alias); |
- if (canonical == NULL) { |
- canonical = Place::Wrap(isolate_, |
- alias, |
- kAnyInstanceAnyIndexAlias + aliases_.length()); |
- InsertAlias(canonical); |
- } |
- return canonical; |
- } |
- |
- BitVector* GetRepresentativesSet(intptr_t alias) { |
- return (alias < representatives_.length()) ? representatives_[alias] : NULL; |
- } |
- |
- BitVector* EnsureSet(GrowableArray<BitVector*>* sets, |
- intptr_t alias) { |
- while (sets->length() <= alias) { |
- sets->Add(NULL); |
- } |
- |
- BitVector* set = (*sets)[alias]; |
- if (set == NULL) { |
- (*sets)[alias] = set = new(isolate_) BitVector(isolate_, max_place_id()); |
- } |
- return set; |
- } |
- |
- void AddAllRepresentatives(const Place* to, intptr_t from) { |
- AddAllRepresentatives(to->id(), from); |
- } |
- |
- void AddAllRepresentatives(intptr_t to, intptr_t from) { |
- BitVector* from_set = GetRepresentativesSet(from); |
- if (from_set != NULL) { |
- EnsureSet(&killed_, to)->AddAll(from_set); |
- } |
- } |
- |
- void CrossAlias(const Place* to, const Place& from) { |
- const intptr_t from_id = LookupAliasId(from); |
- if (from_id == kNoAlias) { |
- return; |
- } |
- CrossAlias(to, from_id); |
- } |
- |
- void CrossAlias(const Place* to, intptr_t from) { |
- AddAllRepresentatives(to->id(), from); |
- AddAllRepresentatives(from, to->id()); |
- } |
- |
- // When computing kill sets we let less generic alias insert its |
- // representatives into more generic alias'es kill set. For example |
- // when visiting alias X[*] instead of searching for all aliases X[C] |
- // and inserting their representatives into kill set for X[*] we update |
- // kill set for X[*] each time we visit new X[C] for some C. |
- // There is an exception however: if both aliases are parametric like *[C] |
- // and X[*] which cross alias when X is an aliased allocation then we use |
- // artificial aliases that contain all possible representatives for the given |
- // alias for any value of the parameter to compute resulting kill set. |
- void ComputeKillSet(const Place* alias) { |
- switch (alias->kind()) { |
- case Place::kIndexed: // Either *[*] or X[*] alias. |
- if (alias->instance() == NULL) { |
- // *[*] aliases with X[*], X[C], *[C]. |
- AddAllRepresentatives(alias, kAnyConstantIndexedAlias); |
- AddAllRepresentatives(alias, kAnyAllocationIndexedAlias); |
- } else if (CanBeAliased(alias->instance())) { |
- // X[*] aliases with X[C]. |
- // If X can be aliased then X[*] also aliases with *[C], *[*]. |
- CrossAlias(alias, kAnyInstanceAnyIndexAlias); |
- AddAllRepresentatives(alias, kUnknownInstanceConstantIndexedAlias); |
- } |
- break; |
- |
- case Place::kConstantIndexed: // Either X[C] or *[C] alias. |
- if (alias->instance() == NULL) { |
- // *[C] aliases with X[C], X[*], *[*]. |
- AddAllRepresentatives(alias, kAnyAllocationIndexedAlias); |
- CrossAlias(alias, kAnyInstanceAnyIndexAlias); |
- } else { |
- // X[C] aliases with X[*]. |
- // If X can be aliased then X[C] also aliases with *[C], *[*]. |
- CrossAlias(alias, alias->CopyWithoutIndex()); |
- if (CanBeAliased(alias->instance())) { |
- CrossAlias(alias, alias->CopyWithoutInstance()); |
- CrossAlias(alias, kAnyInstanceAnyIndexAlias); |
- } |
- } |
- break; |
- |
- case Place::kField: |
- case Place::kVMField: |
- if (CanBeAliased(alias->instance())) { |
- // X.f or X.@offs alias with *.f and *.@offs respectively. |
- CrossAlias(alias, alias->CopyWithoutInstance()); |
- } |
- break; |
- |
- case Place::kNone: |
- UNREACHABLE(); |
- } |
- } |
- |
- // Returns true if the given load is unaffected by external side-effects. |
- // This essentially means that no stores to the same location can |
- // occur in other functions. |
- bool IsIndependentFromEffects(Place* place) { |
- if (place->IsFinalField()) { |
- // Note that we can't use LoadField's is_immutable attribute here because |
- // some VM-fields (those that have no corresponding Field object and |
- // accessed through offset alone) can share offset but have different |
- // immutability properties. |
- // One example is the length property of growable and fixed size list. If |
- // loads of these two properties occur in the same function for the same |
- // receiver then they will get the same expression number. However |
- // immutability of the length of fixed size list does not mean that |
- // growable list also has immutable property. Thus we will make a |
- // conservative assumption for the VM-properties. |
- // TODO(vegorov): disambiguate immutable and non-immutable VM-fields with |
- // the same offset e.g. through recognized kind. |
- return true; |
- } |
- |
- return ((place->kind() == Place::kField) || |
- (place->kind() == Place::kVMField)) && |
- !CanBeAliased(place->instance()); |
- } |
- |
- // Returns true if there are direct loads from the given place. |
- bool HasLoadsFromPlace(Definition* defn, const Place* place) { |
- ASSERT((place->kind() == Place::kField) || |
- (place->kind() == Place::kVMField)); |
- |
- for (Value* use = defn->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- Instruction* instr = use->instruction(); |
- if ((instr->IsRedefinition() || |
- instr->IsAssertAssignable()) && |
- HasLoadsFromPlace(instr->AsDefinition(), place)) { |
- return true; |
- } |
- bool is_load = false, is_store; |
- Place load_place(instr, &is_load, &is_store); |
- |
- if (is_load && load_place.Equals(place)) { |
- return true; |
- } |
- } |
- |
- return false; |
- } |
- |
- // Check if any use of the definition can create an alias. |
- // Can add more objects into aliasing_worklist_. |
- bool AnyUseCreatesAlias(Definition* defn) { |
- for (Value* use = defn->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- Instruction* instr = use->instruction(); |
- if (instr->IsPushArgument() || |
- (instr->IsStoreIndexed() |
- && (use->use_index() == StoreIndexedInstr::kValuePos)) || |
- instr->IsStoreStaticField() || |
- instr->IsPhi()) { |
- return true; |
- } else if ((instr->IsAssertAssignable() || instr->IsRedefinition()) && |
- AnyUseCreatesAlias(instr->AsDefinition())) { |
- return true; |
- } else if ((instr->IsStoreInstanceField() |
- && (use->use_index() != StoreInstanceFieldInstr::kInstancePos))) { |
- ASSERT(use->use_index() == StoreInstanceFieldInstr::kValuePos); |
- // If we store this value into an object that is not aliased itself |
- // and we never load again then the store does not create an alias. |
- StoreInstanceFieldInstr* store = instr->AsStoreInstanceField(); |
- Definition* instance = |
- store->instance()->definition()->OriginalDefinition(); |
- if (Place::IsAllocation(instance) && |
- !instance->Identity().IsAliased()) { |
- bool is_load, is_store; |
- Place store_place(instr, &is_load, &is_store); |
- |
- if (!HasLoadsFromPlace(instance, &store_place)) { |
- // No loads found that match this store. If it is yet unknown if |
- // the object is not aliased then optimistically assume this but |
- // add it to the worklist to check its uses transitively. |
- if (instance->Identity().IsUnknown()) { |
- instance->SetIdentity(AliasIdentity::NotAliased()); |
- aliasing_worklist_.Add(instance); |
- } |
- continue; |
- } |
- } |
- return true; |
- } |
- } |
- return false; |
- } |
- |
- // Mark any value stored into the given object as potentially aliased. |
- void MarkStoredValuesEscaping(Definition* defn) { |
- // Find all stores into this object. |
- for (Value* use = defn->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- if (use->instruction()->IsRedefinition() || |
- use->instruction()->IsAssertAssignable()) { |
- MarkStoredValuesEscaping(use->instruction()->AsDefinition()); |
- continue; |
- } |
- if ((use->use_index() == StoreInstanceFieldInstr::kInstancePos) && |
- use->instruction()->IsStoreInstanceField()) { |
- StoreInstanceFieldInstr* store = |
- use->instruction()->AsStoreInstanceField(); |
- Definition* value = store->value()->definition()->OriginalDefinition(); |
- if (value->Identity().IsNotAliased()) { |
- value->SetIdentity(AliasIdentity::Aliased()); |
- identity_rollback_.Add(value); |
- |
- // Add to worklist to propagate the mark transitively. |
- aliasing_worklist_.Add(value); |
- } |
- } |
- } |
- } |
- |
- // Determine if the given definition can't be aliased. |
- void ComputeAliasing(Definition* alloc) { |
- ASSERT(Place::IsAllocation(alloc)); |
- ASSERT(alloc->Identity().IsUnknown()); |
- ASSERT(aliasing_worklist_.is_empty()); |
- |
- alloc->SetIdentity(AliasIdentity::NotAliased()); |
- aliasing_worklist_.Add(alloc); |
- |
- while (!aliasing_worklist_.is_empty()) { |
- Definition* defn = aliasing_worklist_.RemoveLast(); |
- ASSERT(Place::IsAllocation(defn)); |
- // If the definition in the worklist was optimistically marked as |
- // not-aliased check that optimistic assumption still holds: check if |
- // any of its uses can create an alias. |
- if (!defn->Identity().IsAliased() && AnyUseCreatesAlias(defn)) { |
- defn->SetIdentity(AliasIdentity::Aliased()); |
- identity_rollback_.Add(defn); |
- } |
- |
- // If the allocation site is marked as aliased conservatively mark |
- // any values stored into the object aliased too. |
- if (defn->Identity().IsAliased()) { |
- MarkStoredValuesEscaping(defn); |
- } |
- } |
- } |
- |
- Isolate* isolate_; |
- |
- DirectChainedHashMap<PointerKeyValueTrait<Place> >* places_map_; |
- |
- const ZoneGrowableArray<Place*>& places_; |
- |
- const PhiPlaceMoves* phi_moves_; |
- |
- // A list of all seen aliases and a map that allows looking up canonical |
- // alias object. |
- GrowableArray<const Place*> aliases_; |
- DirectChainedHashMap<PointerKeyValueTrait<const Place> > aliases_map_; |
- |
- // Maps alias id to set of ids of places representing the alias. |
- // Place represents an alias if this alias is least generic alias for |
- // the place. |
- // (see ToAlias for the definition of least generic alias). |
- GrowableArray<BitVector*> representatives_; |
- |
- // Maps alias id to set of ids of places aliased. |
- GrowableArray<BitVector*> killed_; |
- |
- // Set of ids of places that can be affected by side-effects other than |
- // explicit stores (i.e. through calls). |
- BitVector* aliased_by_effects_; |
- |
- // Worklist used during alias analysis. |
- GrowableArray<Definition*> aliasing_worklist_; |
- |
- // List of definitions that had their identity set to Aliased. At the end |
- // of load optimization their identity will be rolled back to Unknown to |
- // avoid treating them as Aliased at later stages without checking first |
- // as optimizations can potentially eliminate instructions leading to |
- // aliasing. |
- GrowableArray<Definition*> identity_rollback_; |
-}; |
- |
- |
-static Definition* GetStoredValue(Instruction* instr) { |
- if (instr->IsStoreIndexed()) { |
- return instr->AsStoreIndexed()->value()->definition(); |
- } |
- |
- StoreInstanceFieldInstr* store_instance_field = instr->AsStoreInstanceField(); |
- if (store_instance_field != NULL) { |
- return store_instance_field->value()->definition(); |
- } |
- |
- StoreStaticFieldInstr* store_static_field = instr->AsStoreStaticField(); |
- if (store_static_field != NULL) { |
- return store_static_field->value()->definition(); |
- } |
- |
- UNREACHABLE(); // Should only be called for supported store instructions. |
- return NULL; |
-} |
- |
- |
-static bool IsPhiDependentPlace(Place* place) { |
- return ((place->kind() == Place::kField) || |
- (place->kind() == Place::kVMField)) && |
- (place->instance() != NULL) && |
- place->instance()->IsPhi(); |
-} |
- |
- |
-// For each place that depends on a phi ensure that equivalent places |
-// corresponding to phi input are numbered and record outgoing phi moves |
-// for each block which establish correspondence between phi dependent place |
-// and phi input's place that is flowing in. |
-static PhiPlaceMoves* ComputePhiMoves( |
- DirectChainedHashMap<PointerKeyValueTrait<Place> >* map, |
- ZoneGrowableArray<Place*>* places) { |
- Isolate* isolate = Isolate::Current(); |
- PhiPlaceMoves* phi_moves = new(isolate) PhiPlaceMoves(); |
- |
- for (intptr_t i = 0; i < places->length(); i++) { |
- Place* place = (*places)[i]; |
- |
- if (IsPhiDependentPlace(place)) { |
- PhiInstr* phi = place->instance()->AsPhi(); |
- BlockEntryInstr* block = phi->GetBlock(); |
- |
- if (FLAG_trace_optimization) { |
- OS::Print("phi dependent place %s\n", place->ToCString()); |
- } |
- |
- Place input_place(*place); |
- for (intptr_t j = 0; j < phi->InputCount(); j++) { |
- input_place.set_instance(phi->InputAt(j)->definition()); |
- |
- Place* result = map->Lookup(&input_place); |
- if (result == NULL) { |
- result = Place::Wrap(isolate, input_place, places->length()); |
- map->Insert(result); |
- places->Add(result); |
- if (FLAG_trace_optimization) { |
- OS::Print(" adding place %s as %" Pd "\n", |
- result->ToCString(), |
- result->id()); |
- } |
- } |
- phi_moves->CreateOutgoingMove(isolate, |
- block->PredecessorAt(j), |
- result->id(), |
- place->id()); |
- } |
- } |
- } |
- |
- return phi_moves; |
-} |
- |
- |
-enum CSEMode { |
- kOptimizeLoads, |
- kOptimizeStores |
-}; |
- |
- |
-static AliasedSet* NumberPlaces( |
- FlowGraph* graph, |
- DirectChainedHashMap<PointerKeyValueTrait<Place> >* map, |
- CSEMode mode) { |
- // Loads representing different expression ids will be collected and |
- // used to build per offset kill sets. |
- Isolate* isolate = graph->isolate(); |
- ZoneGrowableArray<Place*>* places = |
- new(isolate) ZoneGrowableArray<Place*>(10); |
- |
- bool has_loads = false; |
- bool has_stores = false; |
- for (BlockIterator it = graph->reverse_postorder_iterator(); |
- !it.Done(); |
- it.Advance()) { |
- BlockEntryInstr* block = it.Current(); |
- |
- for (ForwardInstructionIterator instr_it(block); |
- !instr_it.Done(); |
- instr_it.Advance()) { |
- Instruction* instr = instr_it.Current(); |
- Place place(instr, &has_loads, &has_stores); |
- if (place.kind() == Place::kNone) { |
- continue; |
- } |
- |
- Place* result = map->Lookup(&place); |
- if (result == NULL) { |
- result = Place::Wrap(isolate, place, places->length()); |
- map->Insert(result); |
- places->Add(result); |
- |
- if (FLAG_trace_optimization) { |
- OS::Print("numbering %s as %" Pd "\n", |
- result->ToCString(), |
- result->id()); |
- } |
- } |
- |
- instr->set_place_id(result->id()); |
- } |
- } |
- |
- if ((mode == kOptimizeLoads) && !has_loads) { |
- return NULL; |
- } |
- if ((mode == kOptimizeStores) && !has_stores) { |
- return NULL; |
- } |
- |
- PhiPlaceMoves* phi_moves = ComputePhiMoves(map, places); |
- |
- // Build aliasing sets mapping aliases to loads. |
- return new(isolate) AliasedSet(isolate, map, places, phi_moves); |
-} |
- |
- |
-class LoadOptimizer : public ValueObject { |
- public: |
- LoadOptimizer(FlowGraph* graph, AliasedSet* aliased_set) |
- : graph_(graph), |
- aliased_set_(aliased_set), |
- in_(graph_->preorder().length()), |
- out_(graph_->preorder().length()), |
- gen_(graph_->preorder().length()), |
- kill_(graph_->preorder().length()), |
- exposed_values_(graph_->preorder().length()), |
- out_values_(graph_->preorder().length()), |
- phis_(5), |
- worklist_(5), |
- congruency_worklist_(6), |
- in_worklist_(NULL), |
- forwarded_(false) { |
- const intptr_t num_blocks = graph_->preorder().length(); |
- for (intptr_t i = 0; i < num_blocks; i++) { |
- out_.Add(NULL); |
- gen_.Add(new(I) BitVector(I, aliased_set_->max_place_id())); |
- kill_.Add(new(I) BitVector(I, aliased_set_->max_place_id())); |
- in_.Add(new(I) BitVector(I, aliased_set_->max_place_id())); |
- |
- exposed_values_.Add(NULL); |
- out_values_.Add(NULL); |
- } |
- } |
- |
- ~LoadOptimizer() { |
- aliased_set_->RollbackAliasedIdentites(); |
- } |
- |
- Isolate* isolate() const { return graph_->isolate(); } |
- |
- static bool OptimizeGraph(FlowGraph* graph) { |
- ASSERT(FLAG_load_cse); |
- if (FLAG_trace_load_optimization) { |
- FlowGraphPrinter::PrintGraph("Before LoadOptimizer", graph); |
- } |
- |
- DirectChainedHashMap<PointerKeyValueTrait<Place> > map; |
- AliasedSet* aliased_set = NumberPlaces(graph, &map, kOptimizeLoads); |
- if ((aliased_set != NULL) && !aliased_set->IsEmpty()) { |
- // If any loads were forwarded return true from Optimize to run load |
- // forwarding again. This will allow to forward chains of loads. |
- // This is especially important for context variables as they are built |
- // as loads from loaded context. |
- // TODO(vegorov): renumber newly discovered congruences during the |
- // forwarding to forward chains without running whole pass twice. |
- LoadOptimizer load_optimizer(graph, aliased_set); |
- return load_optimizer.Optimize(); |
- } |
- return false; |
- } |
- |
- private: |
- bool Optimize() { |
- ComputeInitialSets(); |
- ComputeOutSets(); |
- ComputeOutValues(); |
- if (graph_->is_licm_allowed()) { |
- MarkLoopInvariantLoads(); |
- } |
- ForwardLoads(); |
- EmitPhis(); |
- |
- if (FLAG_trace_load_optimization) { |
- FlowGraphPrinter::PrintGraph("After LoadOptimizer", graph_); |
- } |
- |
- return forwarded_; |
- } |
- |
- // Compute sets of loads generated and killed by each block. |
- // Additionally compute upwards exposed and generated loads for each block. |
- // Exposed loads are those that can be replaced if a corresponding |
- // reaching load will be found. |
- // Loads that are locally redundant will be replaced as we go through |
- // instructions. |
- void ComputeInitialSets() { |
- for (BlockIterator block_it = graph_->reverse_postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- const intptr_t preorder_number = block->preorder_number(); |
- |
- BitVector* kill = kill_[preorder_number]; |
- BitVector* gen = gen_[preorder_number]; |
- |
- ZoneGrowableArray<Definition*>* exposed_values = NULL; |
- ZoneGrowableArray<Definition*>* out_values = NULL; |
- |
- for (ForwardInstructionIterator instr_it(block); |
- !instr_it.Done(); |
- instr_it.Advance()) { |
- Instruction* instr = instr_it.Current(); |
- |
- bool is_load = false, is_store = false; |
- Place place(instr, &is_load, &is_store); |
- |
- BitVector* killed = NULL; |
- if (is_store) { |
- const intptr_t alias_id = |
- aliased_set_->LookupAliasId(place.ToAlias()); |
- if (alias_id != AliasedSet::kNoAlias) { |
- killed = aliased_set_->GetKilledSet(alias_id); |
- } else if (!place.IsFinalField()) { |
- // We encountered unknown alias: this means intrablock load |
- // forwarding refined parameter of this store, for example |
- // |
- // o <- alloc() |
- // a.f <- o |
- // u <- a.f |
- // u.x <- null ;; this store alias is *.x |
- // |
- // after intrablock load forwarding |
- // |
- // o <- alloc() |
- // a.f <- o |
- // o.x <- null ;; this store alias is o.x |
- // |
- // In this case we fallback to using place id recorded in the |
- // instruction that still points to the old place with a more |
- // generic alias. |
- const intptr_t old_alias_id = aliased_set_->LookupAliasId( |
- aliased_set_->places()[instr->place_id()]->ToAlias()); |
- killed = aliased_set_->GetKilledSet(old_alias_id); |
- } |
- |
- if (killed != NULL) { |
- kill->AddAll(killed); |
- // There is no need to clear out_values when clearing GEN set |
- // because only those values that are in the GEN set |
- // will ever be used. |
- gen->RemoveAll(killed); |
- } |
- |
- // Only forward stores to normal arrays, float64, and simd arrays |
- // to loads because other array stores (intXX/uintXX/float32) |
- // may implicitly convert the value stored. |
- StoreIndexedInstr* array_store = instr->AsStoreIndexed(); |
- if ((array_store == NULL) || |
- (array_store->class_id() == kArrayCid) || |
- (array_store->class_id() == kTypedDataFloat64ArrayCid) || |
- (array_store->class_id() == kTypedDataFloat32ArrayCid) || |
- (array_store->class_id() == kTypedDataFloat32x4ArrayCid)) { |
- Place* canonical_place = aliased_set_->LookupCanonical(&place); |
- if (canonical_place != NULL) { |
- // Store has a corresponding numbered place that might have a |
- // load. Try forwarding stored value to it. |
- gen->Add(canonical_place->id()); |
- if (out_values == NULL) out_values = CreateBlockOutValues(); |
- (*out_values)[canonical_place->id()] = GetStoredValue(instr); |
- } |
- } |
- |
- ASSERT(!instr->IsDefinition() || |
- !IsLoadEliminationCandidate(instr->AsDefinition())); |
- continue; |
- } else if (is_load) { |
- // Check if this load needs renumbering because of the intrablock |
- // load forwarding. |
- const Place* canonical = aliased_set_->LookupCanonical(&place); |
- if ((canonical != NULL) && |
- (canonical->id() != instr->AsDefinition()->place_id())) { |
- instr->AsDefinition()->set_place_id(canonical->id()); |
- } |
- } |
- |
- // If instruction has effects then kill all loads affected. |
- if (!instr->Effects().IsNone()) { |
- kill->AddAll(aliased_set_->aliased_by_effects()); |
- // There is no need to clear out_values when removing values from GEN |
- // set because only those values that are in the GEN set |
- // will ever be used. |
- gen->RemoveAll(aliased_set_->aliased_by_effects()); |
- continue; |
- } |
- |
- Definition* defn = instr->AsDefinition(); |
- if (defn == NULL) { |
- continue; |
- } |
- |
- // For object allocation forward initial values of the fields to |
- // subsequent loads. For skip final fields. Final fields are |
- // initialized in constructor that potentially can be not inlined into |
- // the function that we are currently optimizing. However at the same |
- // time we assume that values of the final fields can be forwarded |
- // across side-effects. If we add 'null' as known values for these |
- // fields here we will incorrectly propagate this null across |
- // constructor invocation. |
- AllocateObjectInstr* alloc = instr->AsAllocateObject(); |
- if ((alloc != NULL)) { |
- for (Value* use = alloc->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- // Look for all immediate loads from this object. |
- if (use->use_index() != 0) { |
- continue; |
- } |
- |
- LoadFieldInstr* load = use->instruction()->AsLoadField(); |
- if (load != NULL) { |
- // Found a load. Initialize current value of the field to null for |
- // normal fields, or with type arguments. |
- |
- // Forward for all fields for non-escaping objects and only |
- // non-final fields and type arguments for escaping ones. |
- if (aliased_set_->CanBeAliased(alloc) && |
- (load->field() != NULL) && |
- load->field()->is_final()) { |
- continue; |
- } |
- |
- Definition* forward_def = graph_->constant_null(); |
- if (alloc->ArgumentCount() > 0) { |
- ASSERT(alloc->ArgumentCount() == 1); |
- intptr_t type_args_offset = |
- alloc->cls().type_arguments_field_offset(); |
- if (load->offset_in_bytes() == type_args_offset) { |
- forward_def = alloc->PushArgumentAt(0)->value()->definition(); |
- } |
- } |
- gen->Add(load->place_id()); |
- if (out_values == NULL) out_values = CreateBlockOutValues(); |
- (*out_values)[load->place_id()] = forward_def; |
- } |
- } |
- continue; |
- } |
- |
- if (!IsLoadEliminationCandidate(defn)) { |
- continue; |
- } |
- |
- const intptr_t place_id = defn->place_id(); |
- if (gen->Contains(place_id)) { |
- // This is a locally redundant load. |
- ASSERT((out_values != NULL) && ((*out_values)[place_id] != NULL)); |
- |
- Definition* replacement = (*out_values)[place_id]; |
- EnsureSSATempIndex(graph_, defn, replacement); |
- if (FLAG_trace_optimization) { |
- OS::Print("Replacing load v%" Pd " with v%" Pd "\n", |
- defn->ssa_temp_index(), |
- replacement->ssa_temp_index()); |
- } |
- |
- defn->ReplaceUsesWith(replacement); |
- instr_it.RemoveCurrentFromGraph(); |
- forwarded_ = true; |
- continue; |
- } else if (!kill->Contains(place_id)) { |
- // This is an exposed load: it is the first representative of a |
- // given expression id and it is not killed on the path from |
- // the block entry. |
- if (exposed_values == NULL) { |
- static const intptr_t kMaxExposedValuesInitialSize = 5; |
- exposed_values = new(I) ZoneGrowableArray<Definition*>( |
- Utils::Minimum(kMaxExposedValuesInitialSize, |
- aliased_set_->max_place_id())); |
- } |
- |
- exposed_values->Add(defn); |
- } |
- |
- gen->Add(place_id); |
- |
- if (out_values == NULL) out_values = CreateBlockOutValues(); |
- (*out_values)[place_id] = defn; |
- } |
- |
- exposed_values_[preorder_number] = exposed_values; |
- out_values_[preorder_number] = out_values; |
- } |
- } |
- |
- static void PerformPhiMoves(PhiPlaceMoves::MovesList phi_moves, |
- BitVector* out, |
- BitVector* forwarded_loads) { |
- forwarded_loads->Clear(); |
- |
- for (intptr_t i = 0; i < phi_moves->length(); i++) { |
- const intptr_t from = (*phi_moves)[i].from(); |
- const intptr_t to = (*phi_moves)[i].to(); |
- if (from == to) continue; |
- |
- if (out->Contains(from)) { |
- forwarded_loads->Add(to); |
- } |
- } |
- |
- for (intptr_t i = 0; i < phi_moves->length(); i++) { |
- const intptr_t from = (*phi_moves)[i].from(); |
- const intptr_t to = (*phi_moves)[i].to(); |
- if (from == to) continue; |
- |
- out->Remove(to); |
- } |
- |
- out->AddAll(forwarded_loads); |
- } |
- |
- // Compute OUT sets by propagating them iteratively until fix point |
- // is reached. |
- void ComputeOutSets() { |
- BitVector* temp = new(I) BitVector(I, aliased_set_->max_place_id()); |
- BitVector* forwarded_loads = |
- new(I) BitVector(I, aliased_set_->max_place_id()); |
- BitVector* temp_out = new(I) BitVector(I, aliased_set_->max_place_id()); |
- |
- bool changed = true; |
- while (changed) { |
- changed = false; |
- |
- for (BlockIterator block_it = graph_->reverse_postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- |
- const intptr_t preorder_number = block->preorder_number(); |
- |
- BitVector* block_in = in_[preorder_number]; |
- BitVector* block_out = out_[preorder_number]; |
- BitVector* block_kill = kill_[preorder_number]; |
- BitVector* block_gen = gen_[preorder_number]; |
- |
- // Compute block_in as the intersection of all out(p) where p |
- // is a predecessor of the current block. |
- if (block->IsGraphEntry()) { |
- temp->Clear(); |
- } else { |
- temp->SetAll(); |
- ASSERT(block->PredecessorCount() > 0); |
- for (intptr_t i = 0; i < block->PredecessorCount(); i++) { |
- BlockEntryInstr* pred = block->PredecessorAt(i); |
- BitVector* pred_out = out_[pred->preorder_number()]; |
- if (pred_out == NULL) continue; |
- PhiPlaceMoves::MovesList phi_moves = |
- aliased_set_->phi_moves()->GetOutgoingMoves(pred); |
- if (phi_moves != NULL) { |
- // If there are phi moves, perform intersection with |
- // a copy of pred_out where the phi moves are applied. |
- temp_out->CopyFrom(pred_out); |
- PerformPhiMoves(phi_moves, temp_out, forwarded_loads); |
- pred_out = temp_out; |
- } |
- temp->Intersect(pred_out); |
- } |
- } |
- |
- if (!temp->Equals(*block_in) || (block_out == NULL)) { |
- // If IN set has changed propagate the change to OUT set. |
- block_in->CopyFrom(temp); |
- |
- temp->RemoveAll(block_kill); |
- temp->AddAll(block_gen); |
- |
- if ((block_out == NULL) || !block_out->Equals(*temp)) { |
- if (block_out == NULL) { |
- block_out = out_[preorder_number] = |
- new(I) BitVector(I, aliased_set_->max_place_id()); |
- } |
- block_out->CopyFrom(temp); |
- changed = true; |
- } |
- } |
- } |
- } |
- } |
- |
- // Compute out_values mappings by propagating them in reverse postorder once |
- // through the graph. Generate phis on back edges where eager merge is |
- // impossible. |
- // No replacement is done at this point and thus any out_value[place_id] is |
- // changed at most once: from NULL to an actual value. |
- // When merging incoming loads we might need to create a phi. |
- // These phis are not inserted at the graph immediately because some of them |
- // might become redundant after load forwarding is done. |
- void ComputeOutValues() { |
- GrowableArray<PhiInstr*> pending_phis(5); |
- ZoneGrowableArray<Definition*>* temp_forwarded_values = NULL; |
- |
- for (BlockIterator block_it = graph_->reverse_postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- |
- const bool can_merge_eagerly = CanMergeEagerly(block); |
- |
- const intptr_t preorder_number = block->preorder_number(); |
- |
- ZoneGrowableArray<Definition*>* block_out_values = |
- out_values_[preorder_number]; |
- |
- |
- // If OUT set has changed then we have new values available out of |
- // the block. Compute these values creating phi where necessary. |
- for (BitVector::Iterator it(out_[preorder_number]); |
- !it.Done(); |
- it.Advance()) { |
- const intptr_t place_id = it.Current(); |
- |
- if (block_out_values == NULL) { |
- out_values_[preorder_number] = block_out_values = |
- CreateBlockOutValues(); |
- } |
- |
- if ((*block_out_values)[place_id] == NULL) { |
- ASSERT(block->PredecessorCount() > 0); |
- Definition* in_value = can_merge_eagerly ? |
- MergeIncomingValues(block, place_id) : NULL; |
- if ((in_value == NULL) && |
- (in_[preorder_number]->Contains(place_id))) { |
- PhiInstr* phi = new(I) PhiInstr(block->AsJoinEntry(), |
- block->PredecessorCount()); |
- phi->set_place_id(place_id); |
- pending_phis.Add(phi); |
- in_value = phi; |
- } |
- (*block_out_values)[place_id] = in_value; |
- } |
- } |
- |
- // If the block has outgoing phi moves perform them. Use temporary list |
- // of values to ensure that cyclic moves are performed correctly. |
- PhiPlaceMoves::MovesList phi_moves = |
- aliased_set_->phi_moves()->GetOutgoingMoves(block); |
- if ((phi_moves != NULL) && (block_out_values != NULL)) { |
- if (temp_forwarded_values == NULL) { |
- temp_forwarded_values = CreateBlockOutValues(); |
- } |
- |
- for (intptr_t i = 0; i < phi_moves->length(); i++) { |
- const intptr_t from = (*phi_moves)[i].from(); |
- const intptr_t to = (*phi_moves)[i].to(); |
- if (from == to) continue; |
- |
- (*temp_forwarded_values)[to] = (*block_out_values)[from]; |
- } |
- |
- for (intptr_t i = 0; i < phi_moves->length(); i++) { |
- const intptr_t from = (*phi_moves)[i].from(); |
- const intptr_t to = (*phi_moves)[i].to(); |
- if (from == to) continue; |
- |
- (*block_out_values)[to] = (*temp_forwarded_values)[to]; |
- } |
- } |
- |
- if (FLAG_trace_load_optimization) { |
- OS::Print("B%" Pd "\n", block->block_id()); |
- OS::Print(" IN: "); |
- aliased_set_->PrintSet(in_[preorder_number]); |
- OS::Print("\n"); |
- |
- OS::Print(" KILL: "); |
- aliased_set_->PrintSet(kill_[preorder_number]); |
- OS::Print("\n"); |
- |
- OS::Print(" OUT: "); |
- aliased_set_->PrintSet(out_[preorder_number]); |
- OS::Print("\n"); |
- } |
- } |
- |
- // All blocks were visited. Fill pending phis with inputs |
- // that flow on back edges. |
- for (intptr_t i = 0; i < pending_phis.length(); i++) { |
- FillPhiInputs(pending_phis[i]); |
- } |
- } |
- |
- bool CanMergeEagerly(BlockEntryInstr* block) { |
- for (intptr_t i = 0; i < block->PredecessorCount(); i++) { |
- BlockEntryInstr* pred = block->PredecessorAt(i); |
- if (pred->postorder_number() < block->postorder_number()) { |
- return false; |
- } |
- } |
- return true; |
- } |
- |
- void MarkLoopInvariantLoads() { |
- const ZoneGrowableArray<BlockEntryInstr*>& loop_headers = |
- graph_->LoopHeaders(); |
- |
- ZoneGrowableArray<BitVector*>* invariant_loads = |
- new(I) ZoneGrowableArray<BitVector*>(loop_headers.length()); |
- |
- for (intptr_t i = 0; i < loop_headers.length(); i++) { |
- BlockEntryInstr* header = loop_headers[i]; |
- BlockEntryInstr* pre_header = header->ImmediateDominator(); |
- if (pre_header == NULL) { |
- invariant_loads->Add(NULL); |
- continue; |
- } |
- |
- BitVector* loop_gen = new(I) BitVector(I, aliased_set_->max_place_id()); |
- for (BitVector::Iterator loop_it(header->loop_info()); |
- !loop_it.Done(); |
- loop_it.Advance()) { |
- const intptr_t preorder_number = loop_it.Current(); |
- loop_gen->AddAll(gen_[preorder_number]); |
- } |
- |
- for (BitVector::Iterator loop_it(header->loop_info()); |
- !loop_it.Done(); |
- loop_it.Advance()) { |
- const intptr_t preorder_number = loop_it.Current(); |
- loop_gen->RemoveAll(kill_[preorder_number]); |
- } |
- |
- if (FLAG_trace_optimization) { |
- for (BitVector::Iterator it(loop_gen); !it.Done(); it.Advance()) { |
- OS::Print("place %s is loop invariant for B%" Pd "\n", |
- aliased_set_->places()[it.Current()]->ToCString(), |
- header->block_id()); |
- } |
- } |
- |
- invariant_loads->Add(loop_gen); |
- } |
- |
- graph_->set_loop_invariant_loads(invariant_loads); |
- } |
- |
- // Compute incoming value for the given expression id. |
- // Will create a phi if different values are incoming from multiple |
- // predecessors. |
- Definition* MergeIncomingValues(BlockEntryInstr* block, intptr_t place_id) { |
- // First check if the same value is coming in from all predecessors. |
- static Definition* const kDifferentValuesMarker = |
- reinterpret_cast<Definition*>(-1); |
- Definition* incoming = NULL; |
- for (intptr_t i = 0; i < block->PredecessorCount(); i++) { |
- BlockEntryInstr* pred = block->PredecessorAt(i); |
- ZoneGrowableArray<Definition*>* pred_out_values = |
- out_values_[pred->preorder_number()]; |
- if ((pred_out_values == NULL) || ((*pred_out_values)[place_id] == NULL)) { |
- return NULL; |
- } else if (incoming == NULL) { |
- incoming = (*pred_out_values)[place_id]; |
- } else if (incoming != (*pred_out_values)[place_id]) { |
- incoming = kDifferentValuesMarker; |
- } |
- } |
- |
- if (incoming != kDifferentValuesMarker) { |
- ASSERT(incoming != NULL); |
- return incoming; |
- } |
- |
- // Incoming values are different. Phi is required to merge. |
- PhiInstr* phi = new(I) PhiInstr( |
- block->AsJoinEntry(), block->PredecessorCount()); |
- phi->set_place_id(place_id); |
- FillPhiInputs(phi); |
- return phi; |
- } |
- |
- void FillPhiInputs(PhiInstr* phi) { |
- BlockEntryInstr* block = phi->GetBlock(); |
- const intptr_t place_id = phi->place_id(); |
- |
- for (intptr_t i = 0; i < block->PredecessorCount(); i++) { |
- BlockEntryInstr* pred = block->PredecessorAt(i); |
- ZoneGrowableArray<Definition*>* pred_out_values = |
- out_values_[pred->preorder_number()]; |
- ASSERT((*pred_out_values)[place_id] != NULL); |
- |
- // Sets of outgoing values are not linked into use lists so |
- // they might contain values that were replaced and removed |
- // from the graph by this iteration. |
- // To prevent using them we additionally mark definitions themselves |
- // as replaced and store a pointer to the replacement. |
- Definition* replacement = (*pred_out_values)[place_id]->Replacement(); |
- Value* input = new(I) Value(replacement); |
- phi->SetInputAt(i, input); |
- replacement->AddInputUse(input); |
- } |
- |
- phi->set_ssa_temp_index(graph_->alloc_ssa_temp_index()); |
- phis_.Add(phi); // Postpone phi insertion until after load forwarding. |
- |
- if (FLAG_trace_load_optimization) { |
- OS::Print("created pending phi %s for %s at B%" Pd "\n", |
- phi->ToCString(), |
- aliased_set_->places()[place_id]->ToCString(), |
- block->block_id()); |
- } |
- } |
- |
- // Iterate over basic blocks and replace exposed loads with incoming |
- // values. |
- void ForwardLoads() { |
- for (BlockIterator block_it = graph_->reverse_postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- |
- ZoneGrowableArray<Definition*>* loads = |
- exposed_values_[block->preorder_number()]; |
- if (loads == NULL) continue; // No exposed loads. |
- |
- BitVector* in = in_[block->preorder_number()]; |
- |
- for (intptr_t i = 0; i < loads->length(); i++) { |
- Definition* load = (*loads)[i]; |
- if (!in->Contains(load->place_id())) continue; // No incoming value. |
- |
- Definition* replacement = MergeIncomingValues(block, load->place_id()); |
- ASSERT(replacement != NULL); |
- |
- // Sets of outgoing values are not linked into use lists so |
- // they might contain values that were replace and removed |
- // from the graph by this iteration. |
- // To prevent using them we additionally mark definitions themselves |
- // as replaced and store a pointer to the replacement. |
- replacement = replacement->Replacement(); |
- |
- if (load != replacement) { |
- EnsureSSATempIndex(graph_, load, replacement); |
- |
- if (FLAG_trace_optimization) { |
- OS::Print("Replacing load v%" Pd " with v%" Pd "\n", |
- load->ssa_temp_index(), |
- replacement->ssa_temp_index()); |
- } |
- |
- load->ReplaceUsesWith(replacement); |
- load->RemoveFromGraph(); |
- load->SetReplacement(replacement); |
- forwarded_ = true; |
- } |
- } |
- } |
- } |
- |
- // Check if the given phi take the same value on all code paths. |
- // Eliminate it as redundant if this is the case. |
- // When analyzing phi operands assumes that only generated during |
- // this load phase can be redundant. They can be distinguished because |
- // they are not marked alive. |
- // TODO(vegorov): move this into a separate phase over all phis. |
- bool EliminateRedundantPhi(PhiInstr* phi) { |
- Definition* value = NULL; // Possible value of this phi. |
- |
- worklist_.Clear(); |
- if (in_worklist_ == NULL) { |
- in_worklist_ = new(I) BitVector(I, graph_->current_ssa_temp_index()); |
- } else { |
- in_worklist_->Clear(); |
- } |
- |
- worklist_.Add(phi); |
- in_worklist_->Add(phi->ssa_temp_index()); |
- |
- for (intptr_t i = 0; i < worklist_.length(); i++) { |
- PhiInstr* phi = worklist_[i]; |
- |
- for (intptr_t i = 0; i < phi->InputCount(); i++) { |
- Definition* input = phi->InputAt(i)->definition(); |
- if (input == phi) continue; |
- |
- PhiInstr* phi_input = input->AsPhi(); |
- if ((phi_input != NULL) && !phi_input->is_alive()) { |
- if (!in_worklist_->Contains(phi_input->ssa_temp_index())) { |
- worklist_.Add(phi_input); |
- in_worklist_->Add(phi_input->ssa_temp_index()); |
- } |
- continue; |
- } |
- |
- if (value == NULL) { |
- value = input; |
- } else if (value != input) { |
- return false; // This phi is not redundant. |
- } |
- } |
- } |
- |
- // All phis in the worklist are redundant and have the same computed |
- // value on all code paths. |
- ASSERT(value != NULL); |
- for (intptr_t i = 0; i < worklist_.length(); i++) { |
- worklist_[i]->ReplaceUsesWith(value); |
- } |
- |
- return true; |
- } |
- |
- // Returns true if definitions are congruent assuming their inputs |
- // are congruent. |
- bool CanBeCongruent(Definition* a, Definition* b) { |
- return (a->tag() == b->tag()) && |
- ((a->IsPhi() && (a->GetBlock() == b->GetBlock())) || |
- (a->AllowsCSE() && a->Dependencies().IsNone() && |
- a->AttributesEqual(b))); |
- } |
- |
- // Given two definitions check if they are congruent under assumption that |
- // their inputs will be proven congruent. If they are - add them to the |
- // worklist to check their inputs' congruency. |
- // Returns true if pair was added to the worklist or is already in the |
- // worklist and false if a and b are not congruent. |
- bool AddPairToCongruencyWorklist(Definition* a, Definition* b) { |
- if (!CanBeCongruent(a, b)) { |
- return false; |
- } |
- |
- // If a is already in the worklist check if it is being compared to b. |
- // Give up if it is not. |
- if (in_worklist_->Contains(a->ssa_temp_index())) { |
- for (intptr_t i = 0; i < congruency_worklist_.length(); i += 2) { |
- if (a == congruency_worklist_[i]) { |
- return (b == congruency_worklist_[i + 1]); |
- } |
- } |
- UNREACHABLE(); |
- } else if (in_worklist_->Contains(b->ssa_temp_index())) { |
- return AddPairToCongruencyWorklist(b, a); |
- } |
- |
- congruency_worklist_.Add(a); |
- congruency_worklist_.Add(b); |
- in_worklist_->Add(a->ssa_temp_index()); |
- return true; |
- } |
- |
- bool AreInputsCongruent(Definition* a, Definition* b) { |
- ASSERT(a->tag() == b->tag()); |
- ASSERT(a->InputCount() == b->InputCount()); |
- for (intptr_t j = 0; j < a->InputCount(); j++) { |
- Definition* inputA = a->InputAt(j)->definition(); |
- Definition* inputB = b->InputAt(j)->definition(); |
- |
- if (inputA != inputB) { |
- if (!AddPairToCongruencyWorklist(inputA, inputB)) { |
- return false; |
- } |
- } |
- } |
- return true; |
- } |
- |
- // Returns true if instruction dom dominates instruction other. |
- static bool Dominates(Instruction* dom, Instruction* other) { |
- BlockEntryInstr* dom_block = dom->GetBlock(); |
- BlockEntryInstr* other_block = other->GetBlock(); |
- |
- if (dom_block == other_block) { |
- for (Instruction* current = dom->next(); |
- current != NULL; |
- current = current->next()) { |
- if (current == other) { |
- return true; |
- } |
- } |
- return false; |
- } |
- |
- return dom_block->Dominates(other_block); |
- } |
- |
- // Replace the given phi with another if they are congruent. |
- // Returns true if succeeds. |
- bool ReplacePhiWith(PhiInstr* phi, PhiInstr* replacement) { |
- ASSERT(phi->InputCount() == replacement->InputCount()); |
- ASSERT(phi->block() == replacement->block()); |
- |
- congruency_worklist_.Clear(); |
- if (in_worklist_ == NULL) { |
- in_worklist_ = new(I) BitVector(I, graph_->current_ssa_temp_index()); |
- } else { |
- in_worklist_->Clear(); |
- } |
- |
- // During the comparison worklist contains pairs of definitions to be |
- // compared. |
- if (!AddPairToCongruencyWorklist(phi, replacement)) { |
- return false; |
- } |
- |
- // Process the worklist. It might grow during each comparison step. |
- for (intptr_t i = 0; i < congruency_worklist_.length(); i += 2) { |
- if (!AreInputsCongruent(congruency_worklist_[i], |
- congruency_worklist_[i + 1])) { |
- return false; |
- } |
- } |
- |
- // At this point worklist contains pairs of congruent definitions. |
- // Replace the one member of the pair with another maintaining proper |
- // domination relation between definitions and uses. |
- for (intptr_t i = 0; i < congruency_worklist_.length(); i += 2) { |
- Definition* a = congruency_worklist_[i]; |
- Definition* b = congruency_worklist_[i + 1]; |
- |
- // If these definitions are not phis then we need to pick up one |
- // that dominates another as the replacement: if a dominates b swap them. |
- // Note: both a and b are used as a phi input at the same block B which |
- // means a dominates B and b dominates B, which guarantees that either |
- // a dominates b or b dominates a. |
- if (!a->IsPhi()) { |
- if (Dominates(a, b)) { |
- Definition* t = a; |
- a = b; |
- b = t; |
- } |
- ASSERT(Dominates(b, a)); |
- } |
- |
- if (FLAG_trace_load_optimization) { |
- OS::Print("Replacing %s with congruent %s\n", |
- a->ToCString(), |
- b->ToCString()); |
- } |
- |
- a->ReplaceUsesWith(b); |
- if (a->IsPhi()) { |
- // We might be replacing a phi introduced by the load forwarding |
- // that is not inserted in the graph yet. |
- ASSERT(b->IsPhi()); |
- PhiInstr* phi_a = a->AsPhi(); |
- if (phi_a->is_alive()) { |
- phi_a->mark_dead(); |
- phi_a->block()->RemovePhi(phi_a); |
- phi_a->UnuseAllInputs(); |
- } |
- } else { |
- a->RemoveFromGraph(); |
- } |
- } |
- |
- return true; |
- } |
- |
- // Insert the given phi into the graph. Attempt to find an equal one in the |
- // target block first. |
- // Returns true if the phi was inserted and false if it was replaced. |
- bool EmitPhi(PhiInstr* phi) { |
- for (PhiIterator it(phi->block()); !it.Done(); it.Advance()) { |
- if (ReplacePhiWith(phi, it.Current())) { |
- return false; |
- } |
- } |
- |
- phi->mark_alive(); |
- phi->block()->InsertPhi(phi); |
- return true; |
- } |
- |
- // Phis have not yet been inserted into the graph but they have uses of |
- // their inputs. Insert the non-redundant ones and clear the input uses |
- // of the redundant ones. |
- void EmitPhis() { |
- // First eliminate all redundant phis. |
- for (intptr_t i = 0; i < phis_.length(); i++) { |
- PhiInstr* phi = phis_[i]; |
- if (!phi->HasUses() || EliminateRedundantPhi(phi)) { |
- phi->UnuseAllInputs(); |
- phis_[i] = NULL; |
- } |
- } |
- |
- // Now emit phis or replace them with equal phis already present in the |
- // graph. |
- for (intptr_t i = 0; i < phis_.length(); i++) { |
- PhiInstr* phi = phis_[i]; |
- if ((phi != NULL) && (!phi->HasUses() || !EmitPhi(phi))) { |
- phi->UnuseAllInputs(); |
- } |
- } |
- } |
- |
- ZoneGrowableArray<Definition*>* CreateBlockOutValues() { |
- ZoneGrowableArray<Definition*>* out = |
- new(I) ZoneGrowableArray<Definition*>(aliased_set_->max_place_id()); |
- for (intptr_t i = 0; i < aliased_set_->max_place_id(); i++) { |
- out->Add(NULL); |
- } |
- return out; |
- } |
- |
- FlowGraph* graph_; |
- DirectChainedHashMap<PointerKeyValueTrait<Place> >* map_; |
- |
- // Mapping between field offsets in words and expression ids of loads from |
- // that offset. |
- AliasedSet* aliased_set_; |
- |
- // Per block sets of expression ids for loads that are: incoming (available |
- // on the entry), outgoing (available on the exit), generated and killed. |
- GrowableArray<BitVector*> in_; |
- GrowableArray<BitVector*> out_; |
- GrowableArray<BitVector*> gen_; |
- GrowableArray<BitVector*> kill_; |
- |
- // Per block list of upwards exposed loads. |
- GrowableArray<ZoneGrowableArray<Definition*>*> exposed_values_; |
- |
- // Per block mappings between expression ids and outgoing definitions that |
- // represent those ids. |
- GrowableArray<ZoneGrowableArray<Definition*>*> out_values_; |
- |
- // List of phis generated during ComputeOutValues and ForwardLoads. |
- // Some of these phis might be redundant and thus a separate pass is |
- // needed to emit only non-redundant ones. |
- GrowableArray<PhiInstr*> phis_; |
- |
- // Auxiliary worklist used by redundant phi elimination. |
- GrowableArray<PhiInstr*> worklist_; |
- GrowableArray<Definition*> congruency_worklist_; |
- BitVector* in_worklist_; |
- |
- |
- // True if any load was eliminated. |
- bool forwarded_; |
- |
- DISALLOW_COPY_AND_ASSIGN(LoadOptimizer); |
-}; |
- |
- |
-class StoreOptimizer : public LivenessAnalysis { |
- public: |
- StoreOptimizer(FlowGraph* graph, |
- AliasedSet* aliased_set, |
- DirectChainedHashMap<PointerKeyValueTrait<Place> >* map) |
- : LivenessAnalysis(aliased_set->max_place_id(), graph->postorder()), |
- graph_(graph), |
- map_(map), |
- aliased_set_(aliased_set), |
- exposed_stores_(graph_->postorder().length()) { |
- const intptr_t num_blocks = graph_->postorder().length(); |
- for (intptr_t i = 0; i < num_blocks; i++) { |
- exposed_stores_.Add(NULL); |
- } |
- } |
- |
- static void OptimizeGraph(FlowGraph* graph) { |
- ASSERT(FLAG_load_cse); |
- if (FLAG_trace_load_optimization) { |
- FlowGraphPrinter::PrintGraph("Before StoreOptimizer", graph); |
- } |
- |
- DirectChainedHashMap<PointerKeyValueTrait<Place> > map; |
- AliasedSet* aliased_set = NumberPlaces(graph, &map, kOptimizeStores); |
- if ((aliased_set != NULL) && !aliased_set->IsEmpty()) { |
- StoreOptimizer store_optimizer(graph, aliased_set, &map); |
- store_optimizer.Optimize(); |
- } |
- } |
- |
- private: |
- void Optimize() { |
- Analyze(); |
- if (FLAG_trace_load_optimization) { |
- Dump(); |
- } |
- EliminateDeadStores(); |
- if (FLAG_trace_load_optimization) { |
- FlowGraphPrinter::PrintGraph("After StoreOptimizer", graph_); |
- } |
- } |
- |
- bool CanEliminateStore(Instruction* instr) { |
- switch (instr->tag()) { |
- case Instruction::kStoreInstanceField: |
- if (instr->AsStoreInstanceField()->is_initialization()) { |
- // Can't eliminate stores that initialized unboxed fields. |
- return false; |
- } |
- case Instruction::kStoreIndexed: |
- case Instruction::kStoreStaticField: |
- return true; |
- default: |
- UNREACHABLE(); |
- return false; |
- } |
- } |
- |
- virtual void ComputeInitialSets() { |
- Isolate* isolate = graph_->isolate(); |
- BitVector* all_places = new(isolate) BitVector(isolate, |
- aliased_set_->max_place_id()); |
- all_places->SetAll(); |
- for (BlockIterator block_it = graph_->postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- const intptr_t postorder_number = block->postorder_number(); |
- |
- BitVector* kill = kill_[postorder_number]; |
- BitVector* live_in = live_in_[postorder_number]; |
- BitVector* live_out = live_out_[postorder_number]; |
- |
- ZoneGrowableArray<Instruction*>* exposed_stores = NULL; |
- |
- // Iterate backwards starting at the last instruction. |
- for (BackwardInstructionIterator instr_it(block); |
- !instr_it.Done(); |
- instr_it.Advance()) { |
- Instruction* instr = instr_it.Current(); |
- |
- bool is_load = false; |
- bool is_store = false; |
- Place place(instr, &is_load, &is_store); |
- if (place.IsFinalField()) { |
- // Loads/stores of final fields do not participate. |
- continue; |
- } |
- |
- // Handle stores. |
- if (is_store) { |
- if (kill->Contains(instr->place_id())) { |
- if (!live_in->Contains(instr->place_id()) && |
- CanEliminateStore(instr)) { |
- if (FLAG_trace_optimization) { |
- OS::Print( |
- "Removing dead store to place %" Pd " in block B%" Pd "\n", |
- instr->place_id(), block->block_id()); |
- } |
- instr_it.RemoveCurrentFromGraph(); |
- } |
- } else if (!live_in->Contains(instr->place_id())) { |
- // Mark this store as down-ward exposed: They are the only |
- // candidates for the global store elimination. |
- if (exposed_stores == NULL) { |
- const intptr_t kMaxExposedStoresInitialSize = 5; |
- exposed_stores = new(isolate) ZoneGrowableArray<Instruction*>( |
- Utils::Minimum(kMaxExposedStoresInitialSize, |
- aliased_set_->max_place_id())); |
- } |
- exposed_stores->Add(instr); |
- } |
- // Interfering stores kill only loads from the same place. |
- kill->Add(instr->place_id()); |
- live_in->Remove(instr->place_id()); |
- continue; |
- } |
- |
- // Handle side effects, deoptimization and function return. |
- if (!instr->Effects().IsNone() || |
- instr->CanDeoptimize() || |
- instr->IsThrow() || |
- instr->IsReThrow() || |
- instr->IsReturn()) { |
- // Instructions that return from the function, instructions with side |
- // effects and instructions that can deoptimize are considered as |
- // loads from all places. |
- live_in->CopyFrom(all_places); |
- if (instr->IsThrow() || instr->IsReThrow() || instr->IsReturn()) { |
- // Initialize live-out for exit blocks since it won't be computed |
- // otherwise during the fixed point iteration. |
- live_out->CopyFrom(all_places); |
- } |
- continue; |
- } |
- |
- // Handle loads. |
- Definition* defn = instr->AsDefinition(); |
- if ((defn != NULL) && IsLoadEliminationCandidate(defn)) { |
- const intptr_t alias = aliased_set_->LookupAliasId(place.ToAlias()); |
- live_in->AddAll(aliased_set_->GetKilledSet(alias)); |
- continue; |
- } |
- } |
- exposed_stores_[postorder_number] = exposed_stores; |
- } |
- if (FLAG_trace_load_optimization) { |
- Dump(); |
- OS::Print("---\n"); |
- } |
- } |
- |
- void EliminateDeadStores() { |
- // Iteration order does not matter here. |
- for (BlockIterator block_it = graph_->postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- const intptr_t postorder_number = block->postorder_number(); |
- |
- BitVector* live_out = live_out_[postorder_number]; |
- |
- ZoneGrowableArray<Instruction*>* exposed_stores = |
- exposed_stores_[postorder_number]; |
- if (exposed_stores == NULL) continue; // No exposed stores. |
- |
- // Iterate over candidate stores. |
- for (intptr_t i = 0; i < exposed_stores->length(); ++i) { |
- Instruction* instr = (*exposed_stores)[i]; |
- bool is_load = false; |
- bool is_store = false; |
- Place place(instr, &is_load, &is_store); |
- ASSERT(!is_load && is_store); |
- if (place.IsFinalField()) { |
- // Final field do not participate in dead store elimination. |
- continue; |
- } |
- // Eliminate a downward exposed store if the corresponding place is not |
- // in live-out. |
- if (!live_out->Contains(instr->place_id()) && |
- CanEliminateStore(instr)) { |
- if (FLAG_trace_optimization) { |
- OS::Print("Removing dead store to place %" Pd " block B%" Pd "\n", |
- instr->place_id(), block->block_id()); |
- } |
- instr->RemoveFromGraph(/* ignored */ false); |
- } |
- } |
- } |
- } |
- |
- FlowGraph* graph_; |
- DirectChainedHashMap<PointerKeyValueTrait<Place> >* map_; |
- |
- // Mapping between field offsets in words and expression ids of loads from |
- // that offset. |
- AliasedSet* aliased_set_; |
- |
- // Per block list of downward exposed stores. |
- GrowableArray<ZoneGrowableArray<Instruction*>*> exposed_stores_; |
- |
- DISALLOW_COPY_AND_ASSIGN(StoreOptimizer); |
-}; |
- |
- |
-void DeadStoreElimination::Optimize(FlowGraph* graph) { |
- if (FLAG_dead_store_elimination) { |
- StoreOptimizer::OptimizeGraph(graph); |
- } |
-} |
- |
- |
-// Returns true iff this definition is used in a non-phi instruction. |
-static bool HasRealUse(Definition* def) { |
- // Environment uses are real (non-phi) uses. |
- if (def->env_use_list() != NULL) return true; |
- |
- for (Value::Iterator it(def->input_use_list()); |
- !it.Done(); |
- it.Advance()) { |
- if (!it.Current()->instruction()->IsPhi()) return true; |
- } |
- return false; |
-} |
- |
- |
-void DeadCodeElimination::EliminateDeadPhis(FlowGraph* flow_graph) { |
- GrowableArray<PhiInstr*> live_phis; |
- for (BlockIterator b = flow_graph->postorder_iterator(); |
- !b.Done(); |
- b.Advance()) { |
- JoinEntryInstr* join = b.Current()->AsJoinEntry(); |
- if (join != NULL) { |
- for (PhiIterator it(join); !it.Done(); it.Advance()) { |
- PhiInstr* phi = it.Current(); |
- // Phis that have uses and phis inside try blocks are |
- // marked as live. |
- if (HasRealUse(phi) || join->InsideTryBlock()) { |
- live_phis.Add(phi); |
- phi->mark_alive(); |
- } else { |
- phi->mark_dead(); |
- } |
- } |
- } |
- } |
- |
- while (!live_phis.is_empty()) { |
- PhiInstr* phi = live_phis.RemoveLast(); |
- for (intptr_t i = 0; i < phi->InputCount(); i++) { |
- Value* val = phi->InputAt(i); |
- PhiInstr* used_phi = val->definition()->AsPhi(); |
- if ((used_phi != NULL) && !used_phi->is_alive()) { |
- used_phi->mark_alive(); |
- live_phis.Add(used_phi); |
- } |
- } |
- } |
- |
- for (BlockIterator it(flow_graph->postorder_iterator()); |
- !it.Done(); |
- it.Advance()) { |
- JoinEntryInstr* join = it.Current()->AsJoinEntry(); |
- if (join != NULL) { |
- if (join->phis_ == NULL) continue; |
- |
- // Eliminate dead phis and compact the phis_ array of the block. |
- intptr_t to_index = 0; |
- for (intptr_t i = 0; i < join->phis_->length(); ++i) { |
- PhiInstr* phi = (*join->phis_)[i]; |
- if (phi != NULL) { |
- if (!phi->is_alive()) { |
- phi->ReplaceUsesWith(flow_graph->constant_null()); |
- phi->UnuseAllInputs(); |
- (*join->phis_)[i] = NULL; |
- if (FLAG_trace_optimization) { |
- OS::Print("Removing dead phi v%" Pd "\n", phi->ssa_temp_index()); |
- } |
- } else if (phi->IsRedundant()) { |
- phi->ReplaceUsesWith(phi->InputAt(0)->definition()); |
- phi->UnuseAllInputs(); |
- (*join->phis_)[i] = NULL; |
- if (FLAG_trace_optimization) { |
- OS::Print("Removing redundant phi v%" Pd "\n", |
- phi->ssa_temp_index()); |
- } |
- } else { |
- (*join->phis_)[to_index++] = phi; |
- } |
- } |
- } |
- if (to_index == 0) { |
- join->phis_ = NULL; |
- } else { |
- join->phis_->TruncateTo(to_index); |
- } |
- } |
- } |
-} |
- |
- |
-class CSEInstructionMap : public ValueObject { |
- public: |
- // Right now CSE and LICM track a single effect: possible externalization of |
- // strings. |
- // Other effects like modifications of fields are tracked in a separate load |
- // forwarding pass via Alias structure. |
- COMPILE_ASSERT(EffectSet::kLastEffect == 1); |
- |
- CSEInstructionMap() : independent_(), dependent_() { } |
- explicit CSEInstructionMap(const CSEInstructionMap& other) |
- : ValueObject(), |
- independent_(other.independent_), |
- dependent_(other.dependent_) { |
- } |
- |
- void RemoveAffected(EffectSet effects) { |
- if (!effects.IsNone()) { |
- dependent_.Clear(); |
- } |
- } |
- |
- Instruction* Lookup(Instruction* other) const { |
- return GetMapFor(other)->Lookup(other); |
- } |
- |
- void Insert(Instruction* instr) { |
- return GetMapFor(instr)->Insert(instr); |
- } |
- |
- private: |
- typedef DirectChainedHashMap<PointerKeyValueTrait<Instruction> > Map; |
- |
- Map* GetMapFor(Instruction* instr) { |
- return instr->Dependencies().IsNone() ? &independent_ : &dependent_; |
- } |
- |
- const Map* GetMapFor(Instruction* instr) const { |
- return instr->Dependencies().IsNone() ? &independent_ : &dependent_; |
- } |
- |
- // All computations that are not affected by any side-effect. |
- // Majority of computations are not affected by anything and will be in |
- // this map. |
- Map independent_; |
- |
- // All computations that are affected by side effect. |
- Map dependent_; |
-}; |
- |
- |
-bool DominatorBasedCSE::Optimize(FlowGraph* graph) { |
- bool changed = false; |
- if (FLAG_load_cse) { |
- changed = LoadOptimizer::OptimizeGraph(graph) || changed; |
- } |
- |
- CSEInstructionMap map; |
- changed = OptimizeRecursive(graph, graph->graph_entry(), &map) || changed; |
- |
- return changed; |
-} |
- |
- |
-bool DominatorBasedCSE::OptimizeRecursive( |
- FlowGraph* graph, |
- BlockEntryInstr* block, |
- CSEInstructionMap* map) { |
- bool changed = false; |
- for (ForwardInstructionIterator it(block); !it.Done(); it.Advance()) { |
- Instruction* current = it.Current(); |
- if (current->AllowsCSE()) { |
- Instruction* replacement = map->Lookup(current); |
- if ((replacement != NULL) && |
- graph->block_effects()->IsAvailableAt(replacement, block)) { |
- // Replace current with lookup result. |
- ReplaceCurrentInstruction(&it, current, replacement, graph); |
- changed = true; |
- continue; |
- } |
- |
- // For simplicity we assume that instruction either does not depend on |
- // anything or does not affect anything. If this is not the case then |
- // we should first remove affected instructions from the map and |
- // then add instruction to the map so that it does not kill itself. |
- ASSERT(current->Effects().IsNone() || current->Dependencies().IsNone()); |
- map->Insert(current); |
- } |
- |
- map->RemoveAffected(current->Effects()); |
- } |
- |
- // Process children in the dominator tree recursively. |
- intptr_t num_children = block->dominated_blocks().length(); |
- for (intptr_t i = 0; i < num_children; ++i) { |
- BlockEntryInstr* child = block->dominated_blocks()[i]; |
- if (i < num_children - 1) { |
- // Copy map. |
- CSEInstructionMap child_map(*map); |
- changed = OptimizeRecursive(graph, child, &child_map) || changed; |
- } else { |
- // Reuse map for the last child. |
- changed = OptimizeRecursive(graph, child, map) || changed; |
- } |
- } |
- return changed; |
-} |
- |
- |
ConstantPropagator::ConstantPropagator( |
FlowGraph* graph, |
const GrowableArray<BlockEntryInstr*>& ignored) |
@@ -8202,7 +725,7 @@ |
if ((checked_type.IsFloat32x4Type() && (rep == kUnboxedFloat32x4)) || |
(checked_type.IsInt32x4Type() && (rep == kUnboxedInt32x4)) || |
(checked_type.IsDoubleType() && (rep == kUnboxedDouble) && |
- CanUnboxDouble()) || |
+ FlowGraphCompiler::SupportsUnboxedDoubles()) || |
(checked_type.IsIntType() && (rep == kUnboxedMint))) { |
// Ensure that compile time type matches representation. |
ASSERT(((rep == kUnboxedFloat32x4) && (value_cid == kFloat32x4Cid)) || |
@@ -9112,976 +1635,4 @@ |
} |
} |
- |
-// Returns true if the given phi has a single input use and |
-// is used in the environments either at the corresponding block entry or |
-// at the same instruction where input use is. |
-static bool PhiHasSingleUse(PhiInstr* phi, Value* use) { |
- if ((use->next_use() != NULL) || (phi->input_use_list() != use)) { |
- return false; |
- } |
- |
- BlockEntryInstr* block = phi->block(); |
- for (Value* env_use = phi->env_use_list(); |
- env_use != NULL; |
- env_use = env_use->next_use()) { |
- if ((env_use->instruction() != block) && |
- (env_use->instruction() != use->instruction())) { |
- return false; |
- } |
- } |
- |
- return true; |
-} |
- |
- |
-bool BranchSimplifier::Match(JoinEntryInstr* block) { |
- // Match the pattern of a branch on a comparison whose left operand is a |
- // phi from the same block, and whose right operand is a constant. |
- // |
- // Branch(Comparison(kind, Phi, Constant)) |
- // |
- // These are the branches produced by inlining in a test context. Also, |
- // the phi has no other uses so they can simply be eliminated. The block |
- // has no other phis and no instructions intervening between the phi and |
- // branch so the block can simply be eliminated. |
- BranchInstr* branch = block->last_instruction()->AsBranch(); |
- ASSERT(branch != NULL); |
- ComparisonInstr* comparison = branch->comparison(); |
- Value* left = comparison->left(); |
- PhiInstr* phi = left->definition()->AsPhi(); |
- Value* right = comparison->right(); |
- ConstantInstr* constant = |
- (right == NULL) ? NULL : right->definition()->AsConstant(); |
- return (phi != NULL) && |
- (constant != NULL) && |
- (phi->GetBlock() == block) && |
- PhiHasSingleUse(phi, left) && |
- (block->next() == branch) && |
- (block->phis()->length() == 1); |
-} |
- |
- |
-JoinEntryInstr* BranchSimplifier::ToJoinEntry(Isolate* isolate, |
- TargetEntryInstr* target) { |
- // Convert a target block into a join block. Branches will be duplicated |
- // so the former true and false targets become joins of the control flows |
- // from all the duplicated branches. |
- JoinEntryInstr* join = |
- new(isolate) JoinEntryInstr(target->block_id(), target->try_index()); |
- join->InheritDeoptTarget(isolate, target); |
- join->LinkTo(target->next()); |
- join->set_last_instruction(target->last_instruction()); |
- target->UnuseAllInputs(); |
- return join; |
-} |
- |
- |
-BranchInstr* BranchSimplifier::CloneBranch(Isolate* isolate, |
- BranchInstr* branch, |
- Value* new_left, |
- Value* new_right) { |
- ComparisonInstr* comparison = branch->comparison(); |
- ComparisonInstr* new_comparison = |
- comparison->CopyWithNewOperands(new_left, new_right); |
- BranchInstr* new_branch = new(isolate) BranchInstr(new_comparison); |
- new_branch->set_is_checked(branch->is_checked()); |
- return new_branch; |
-} |
- |
- |
-void BranchSimplifier::Simplify(FlowGraph* flow_graph) { |
- // Optimize some branches that test the value of a phi. When it is safe |
- // to do so, push the branch to each of the predecessor blocks. This is |
- // an optimization when (a) it can avoid materializing a boolean object at |
- // the phi only to test its value, and (b) it can expose opportunities for |
- // constant propagation and unreachable code elimination. This |
- // optimization is intended to run after inlining which creates |
- // opportunities for optimization (a) and before constant folding which |
- // can perform optimization (b). |
- |
- // Begin with a worklist of join blocks ending in branches. They are |
- // candidates for the pattern below. |
- Isolate* isolate = flow_graph->isolate(); |
- const GrowableArray<BlockEntryInstr*>& postorder = flow_graph->postorder(); |
- GrowableArray<BlockEntryInstr*> worklist(postorder.length()); |
- for (BlockIterator it(postorder); !it.Done(); it.Advance()) { |
- BlockEntryInstr* block = it.Current(); |
- if (block->IsJoinEntry() && block->last_instruction()->IsBranch()) { |
- worklist.Add(block); |
- } |
- } |
- |
- // Rewrite until no more instance of the pattern exists. |
- bool changed = false; |
- while (!worklist.is_empty()) { |
- // All blocks in the worklist are join blocks (ending with a branch). |
- JoinEntryInstr* block = worklist.RemoveLast()->AsJoinEntry(); |
- ASSERT(block != NULL); |
- |
- if (Match(block)) { |
- changed = true; |
- |
- // The branch will be copied and pushed to all the join's |
- // predecessors. Convert the true and false target blocks into join |
- // blocks to join the control flows from all of the true |
- // (respectively, false) targets of the copied branches. |
- // |
- // The converted join block will have no phis, so it cannot be another |
- // instance of the pattern. There is thus no need to add it to the |
- // worklist. |
- BranchInstr* branch = block->last_instruction()->AsBranch(); |
- ASSERT(branch != NULL); |
- JoinEntryInstr* join_true = |
- ToJoinEntry(isolate, branch->true_successor()); |
- JoinEntryInstr* join_false = |
- ToJoinEntry(isolate, branch->false_successor()); |
- |
- ComparisonInstr* comparison = branch->comparison(); |
- PhiInstr* phi = comparison->left()->definition()->AsPhi(); |
- ConstantInstr* constant = comparison->right()->definition()->AsConstant(); |
- ASSERT(constant != NULL); |
- // Copy the constant and branch and push it to all the predecessors. |
- for (intptr_t i = 0, count = block->PredecessorCount(); i < count; ++i) { |
- GotoInstr* old_goto = |
- block->PredecessorAt(i)->last_instruction()->AsGoto(); |
- ASSERT(old_goto != NULL); |
- |
- // Replace the goto in each predecessor with a rewritten branch, |
- // rewritten to use the corresponding phi input instead of the phi. |
- Value* new_left = phi->InputAt(i)->Copy(isolate); |
- Value* new_right = new(isolate) Value(constant); |
- BranchInstr* new_branch = |
- CloneBranch(isolate, branch, new_left, new_right); |
- if (branch->env() == NULL) { |
- new_branch->InheritDeoptTarget(isolate, old_goto); |
- } else { |
- // Take the environment from the branch if it has one. |
- new_branch->InheritDeoptTarget(isolate, branch); |
- // InheritDeoptTarget gave the new branch's comparison the same |
- // deopt id that it gave the new branch. The id should be the |
- // deopt id of the original comparison. |
- new_branch->comparison()->SetDeoptId(*comparison); |
- // The phi can be used in the branch's environment. Rename such |
- // uses. |
- for (Environment::DeepIterator it(new_branch->env()); |
- !it.Done(); |
- it.Advance()) { |
- Value* use = it.CurrentValue(); |
- if (use->definition() == phi) { |
- Definition* replacement = phi->InputAt(i)->definition(); |
- use->RemoveFromUseList(); |
- use->set_definition(replacement); |
- replacement->AddEnvUse(use); |
- } |
- } |
- } |
- |
- new_branch->InsertBefore(old_goto); |
- new_branch->set_next(NULL); // Detaching the goto from the graph. |
- old_goto->UnuseAllInputs(); |
- |
- // Update the predecessor block. We may have created another |
- // instance of the pattern so add it to the worklist if necessary. |
- BlockEntryInstr* branch_block = new_branch->GetBlock(); |
- branch_block->set_last_instruction(new_branch); |
- if (branch_block->IsJoinEntry()) worklist.Add(branch_block); |
- |
- // Connect the branch to the true and false joins, via empty target |
- // blocks. |
- TargetEntryInstr* true_target = |
- new(isolate) TargetEntryInstr(flow_graph->max_block_id() + 1, |
- block->try_index()); |
- true_target->InheritDeoptTarget(isolate, join_true); |
- TargetEntryInstr* false_target = |
- new(isolate) TargetEntryInstr(flow_graph->max_block_id() + 2, |
- block->try_index()); |
- false_target->InheritDeoptTarget(isolate, join_false); |
- flow_graph->set_max_block_id(flow_graph->max_block_id() + 2); |
- *new_branch->true_successor_address() = true_target; |
- *new_branch->false_successor_address() = false_target; |
- GotoInstr* goto_true = new(isolate) GotoInstr(join_true); |
- goto_true->InheritDeoptTarget(isolate, join_true); |
- true_target->LinkTo(goto_true); |
- true_target->set_last_instruction(goto_true); |
- GotoInstr* goto_false = new(isolate) GotoInstr(join_false); |
- goto_false->InheritDeoptTarget(isolate, join_false); |
- false_target->LinkTo(goto_false); |
- false_target->set_last_instruction(goto_false); |
- } |
- // When all predecessors have been rewritten, the original block is |
- // unreachable from the graph. |
- phi->UnuseAllInputs(); |
- branch->UnuseAllInputs(); |
- block->UnuseAllInputs(); |
- ASSERT(!phi->HasUses()); |
- } |
- } |
- |
- if (changed) { |
- // We may have changed the block order and the dominator tree. |
- flow_graph->DiscoverBlocks(); |
- GrowableArray<BitVector*> dominance_frontier; |
- flow_graph->ComputeDominators(&dominance_frontier); |
- } |
-} |
- |
- |
-static bool IsTrivialBlock(BlockEntryInstr* block, Definition* defn) { |
- return (block->IsTargetEntry() && (block->PredecessorCount() == 1)) && |
- ((block->next() == block->last_instruction()) || |
- ((block->next() == defn) && (defn->next() == block->last_instruction()))); |
-} |
- |
- |
-static void EliminateTrivialBlock(BlockEntryInstr* block, |
- Definition* instr, |
- IfThenElseInstr* before) { |
- block->UnuseAllInputs(); |
- block->last_instruction()->UnuseAllInputs(); |
- |
- if ((block->next() == instr) && |
- (instr->next() == block->last_instruction())) { |
- before->previous()->LinkTo(instr); |
- instr->LinkTo(before); |
- } |
-} |
- |
- |
-void IfConverter::Simplify(FlowGraph* flow_graph) { |
- Isolate* isolate = flow_graph->isolate(); |
- bool changed = false; |
- |
- const GrowableArray<BlockEntryInstr*>& postorder = flow_graph->postorder(); |
- for (BlockIterator it(postorder); !it.Done(); it.Advance()) { |
- BlockEntryInstr* block = it.Current(); |
- JoinEntryInstr* join = block->AsJoinEntry(); |
- |
- // Detect diamond control flow pattern which materializes a value depending |
- // on the result of the comparison: |
- // |
- // B_pred: |
- // ... |
- // Branch if COMP goto (B_pred1, B_pred2) |
- // B_pred1: -- trivial block that contains at most one definition |
- // v1 = Constant(...) |
- // goto B_block |
- // B_pred2: -- trivial block that contains at most one definition |
- // v2 = Constant(...) |
- // goto B_block |
- // B_block: |
- // v3 = phi(v1, v2) -- single phi |
- // |
- // and replace it with |
- // |
- // Ba: |
- // v3 = IfThenElse(COMP ? v1 : v2) |
- // |
- if ((join != NULL) && |
- (join->phis() != NULL) && |
- (join->phis()->length() == 1) && |
- (block->PredecessorCount() == 2)) { |
- BlockEntryInstr* pred1 = block->PredecessorAt(0); |
- BlockEntryInstr* pred2 = block->PredecessorAt(1); |
- |
- PhiInstr* phi = (*join->phis())[0]; |
- Value* v1 = phi->InputAt(0); |
- Value* v2 = phi->InputAt(1); |
- |
- if (IsTrivialBlock(pred1, v1->definition()) && |
- IsTrivialBlock(pred2, v2->definition()) && |
- (pred1->PredecessorAt(0) == pred2->PredecessorAt(0))) { |
- BlockEntryInstr* pred = pred1->PredecessorAt(0); |
- BranchInstr* branch = pred->last_instruction()->AsBranch(); |
- ComparisonInstr* comparison = branch->comparison(); |
- |
- // Check if the platform supports efficient branchless IfThenElseInstr |
- // for the given combination of comparison and values flowing from |
- // false and true paths. |
- if (IfThenElseInstr::Supports(comparison, v1, v2)) { |
- Value* if_true = (pred1 == branch->true_successor()) ? v1 : v2; |
- Value* if_false = (pred2 == branch->true_successor()) ? v1 : v2; |
- |
- ComparisonInstr* new_comparison = |
- comparison->CopyWithNewOperands( |
- comparison->left()->Copy(isolate), |
- comparison->right()->Copy(isolate)); |
- IfThenElseInstr* if_then_else = new(isolate) IfThenElseInstr( |
- new_comparison, |
- if_true->Copy(isolate), |
- if_false->Copy(isolate)); |
- flow_graph->InsertBefore(branch, |
- if_then_else, |
- NULL, |
- FlowGraph::kValue); |
- |
- phi->ReplaceUsesWith(if_then_else); |
- |
- // Connect IfThenElseInstr to the first instruction in the merge block |
- // effectively eliminating diamond control flow. |
- // Current block as well as pred1 and pred2 blocks are no longer in |
- // the graph at this point. |
- if_then_else->LinkTo(join->next()); |
- pred->set_last_instruction(join->last_instruction()); |
- |
- // Resulting block must inherit block id from the eliminated current |
- // block to guarantee that ordering of phi operands in its successor |
- // stays consistent. |
- pred->set_block_id(block->block_id()); |
- |
- // If v1 and v2 were defined inside eliminated blocks pred1/pred2 |
- // move them out to the place before inserted IfThenElse instruction. |
- EliminateTrivialBlock(pred1, v1->definition(), if_then_else); |
- EliminateTrivialBlock(pred2, v2->definition(), if_then_else); |
- |
- // Update use lists to reflect changes in the graph. |
- phi->UnuseAllInputs(); |
- branch->UnuseAllInputs(); |
- block->UnuseAllInputs(); |
- |
- // The graph has changed. Recompute dominators and block orders after |
- // this pass is finished. |
- changed = true; |
- } |
- } |
- } |
- } |
- |
- if (changed) { |
- // We may have changed the block order and the dominator tree. |
- flow_graph->DiscoverBlocks(); |
- GrowableArray<BitVector*> dominance_frontier; |
- flow_graph->ComputeDominators(&dominance_frontier); |
- } |
-} |
- |
- |
-void FlowGraphOptimizer::EliminateEnvironments() { |
- // After this pass we can no longer perform LICM and hoist instructions |
- // that can deoptimize. |
- |
- flow_graph_->disallow_licm(); |
- for (intptr_t i = 0; i < block_order_.length(); ++i) { |
- BlockEntryInstr* block = block_order_[i]; |
- block->RemoveEnvironment(); |
- for (ForwardInstructionIterator it(block); !it.Done(); it.Advance()) { |
- Instruction* current = it.Current(); |
- if (!current->CanDeoptimize()) { |
- // TODO(srdjan): --source-lines needs deopt environments to get at |
- // the code for this instruction, however, leaving the environment |
- // changes code. |
- current->RemoveEnvironment(); |
- } |
- } |
- } |
-} |
- |
- |
-enum SafeUseCheck { kOptimisticCheck, kStrictCheck }; |
- |
-// Check if the use is safe for allocation sinking. Allocation sinking |
-// candidates can only be used at store instructions: |
-// |
-// - any store into the allocation candidate itself is unconditionally safe |
-// as it just changes the rematerialization state of this candidate; |
-// - store into another object is only safe if another object is allocation |
-// candidate. |
-// |
-// We use a simple fix-point algorithm to discover the set of valid candidates |
-// (see CollectCandidates method), that's why this IsSafeUse can operate in two |
-// modes: |
-// |
-// - optimistic, when every allocation is assumed to be an allocation |
-// sinking candidate; |
-// - strict, when only marked allocations are assumed to be allocation |
-// sinking candidates. |
-// |
-// Fix-point algorithm in CollectCandiates first collects a set of allocations |
-// optimistically and then checks each collected candidate strictly and unmarks |
-// invalid candidates transitively until only strictly valid ones remain. |
-static bool IsSafeUse(Value* use, SafeUseCheck check_type) { |
- if (use->instruction()->IsMaterializeObject()) { |
- return true; |
- } |
- |
- StoreInstanceFieldInstr* store = use->instruction()->AsStoreInstanceField(); |
- if (store != NULL) { |
- if (use == store->value()) { |
- Definition* instance = store->instance()->definition(); |
- return instance->IsAllocateObject() && |
- ((check_type == kOptimisticCheck) || |
- instance->Identity().IsAllocationSinkingCandidate()); |
- } |
- return true; |
- } |
- |
- return false; |
-} |
- |
- |
-// Right now we are attempting to sink allocation only into |
-// deoptimization exit. So candidate should only be used in StoreInstanceField |
-// instructions that write into fields of the allocated object. |
-// We do not support materialization of the object that has type arguments. |
-static bool IsAllocationSinkingCandidate(Definition* alloc, |
- SafeUseCheck check_type) { |
- for (Value* use = alloc->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- if (!IsSafeUse(use, check_type)) { |
- if (FLAG_trace_optimization) { |
- OS::Print("use of %s at %s is unsafe for allocation sinking\n", |
- alloc->ToCString(), |
- use->instruction()->ToCString()); |
- } |
- return false; |
- } |
- } |
- |
- return true; |
-} |
- |
- |
-// If the given use is a store into an object then return an object we are |
-// storing into. |
-static Definition* StoreInto(Value* use) { |
- StoreInstanceFieldInstr* store = use->instruction()->AsStoreInstanceField(); |
- if (store != NULL) { |
- return store->instance()->definition(); |
- } |
- |
- return NULL; |
-} |
- |
- |
-// Remove the given allocation from the graph. It is not observable. |
-// If deoptimization occurs the object will be materialized. |
-void AllocationSinking::EliminateAllocation(Definition* alloc) { |
- ASSERT(IsAllocationSinkingCandidate(alloc, kStrictCheck)); |
- |
- if (FLAG_trace_optimization) { |
- OS::Print("removing allocation from the graph: v%" Pd "\n", |
- alloc->ssa_temp_index()); |
- } |
- |
- // As an allocation sinking candidate it is only used in stores to its own |
- // fields. Remove these stores. |
- for (Value* use = alloc->input_use_list(); |
- use != NULL; |
- use = alloc->input_use_list()) { |
- use->instruction()->RemoveFromGraph(); |
- } |
- |
- // There should be no environment uses. The pass replaced them with |
- // MaterializeObject instructions. |
-#ifdef DEBUG |
- for (Value* use = alloc->env_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- ASSERT(use->instruction()->IsMaterializeObject()); |
- } |
-#endif |
- ASSERT(alloc->input_use_list() == NULL); |
- alloc->RemoveFromGraph(); |
- if (alloc->ArgumentCount() > 0) { |
- ASSERT(alloc->ArgumentCount() == 1); |
- for (intptr_t i = 0; i < alloc->ArgumentCount(); ++i) { |
- alloc->PushArgumentAt(i)->RemoveFromGraph(); |
- } |
- } |
-} |
- |
- |
-// Find allocation instructions that can be potentially eliminated and |
-// rematerialized at deoptimization exits if needed. See IsSafeUse |
-// for the description of algorithm used below. |
-void AllocationSinking::CollectCandidates() { |
- // Optimistically collect all potential candidates. |
- for (BlockIterator block_it = flow_graph_->reverse_postorder_iterator(); |
- !block_it.Done(); |
- block_it.Advance()) { |
- BlockEntryInstr* block = block_it.Current(); |
- for (ForwardInstructionIterator it(block); !it.Done(); it.Advance()) { |
- { AllocateObjectInstr* alloc = it.Current()->AsAllocateObject(); |
- if ((alloc != NULL) && |
- IsAllocationSinkingCandidate(alloc, kOptimisticCheck)) { |
- alloc->SetIdentity(AliasIdentity::AllocationSinkingCandidate()); |
- candidates_.Add(alloc); |
- } |
- } |
- { AllocateUninitializedContextInstr* alloc = |
- it.Current()->AsAllocateUninitializedContext(); |
- if ((alloc != NULL) && |
- IsAllocationSinkingCandidate(alloc, kOptimisticCheck)) { |
- alloc->SetIdentity(AliasIdentity::AllocationSinkingCandidate()); |
- candidates_.Add(alloc); |
- } |
- } |
- } |
- } |
- |
- // Transitively unmark all candidates that are not strictly valid. |
- bool changed; |
- do { |
- changed = false; |
- for (intptr_t i = 0; i < candidates_.length(); i++) { |
- Definition* alloc = candidates_[i]; |
- if (alloc->Identity().IsAllocationSinkingCandidate()) { |
- if (!IsAllocationSinkingCandidate(alloc, kStrictCheck)) { |
- alloc->SetIdentity(AliasIdentity::Unknown()); |
- changed = true; |
- } |
- } |
- } |
- } while (changed); |
- |
- // Shrink the list of candidates removing all unmarked ones. |
- intptr_t j = 0; |
- for (intptr_t i = 0; i < candidates_.length(); i++) { |
- Definition* alloc = candidates_[i]; |
- if (alloc->Identity().IsAllocationSinkingCandidate()) { |
- if (FLAG_trace_optimization) { |
- OS::Print("discovered allocation sinking candidate: v%" Pd "\n", |
- alloc->ssa_temp_index()); |
- } |
- |
- if (j != i) { |
- candidates_[j] = alloc; |
- } |
- j++; |
- } |
- } |
- candidates_.TruncateTo(j); |
-} |
- |
- |
-// If materialization references an allocation sinking candidate then replace |
-// this reference with a materialization which should have been computed for |
-// this side-exit. CollectAllExits should have collected this exit. |
-void AllocationSinking::NormalizeMaterializations() { |
- for (intptr_t i = 0; i < candidates_.length(); i++) { |
- Definition* alloc = candidates_[i]; |
- |
- Value* next_use; |
- for (Value* use = alloc->input_use_list(); |
- use != NULL; |
- use = next_use) { |
- next_use = use->next_use(); |
- if (use->instruction()->IsMaterializeObject()) { |
- use->BindTo(MaterializationFor(alloc, use->instruction())); |
- } |
- } |
- } |
-} |
- |
- |
-// We transitively insert materializations at each deoptimization exit that |
-// might see the given allocation (see ExitsCollector). Some of this |
-// materializations are not actually used and some fail to compute because |
-// they are inserted in the block that is not dominated by the allocation. |
-// Remove them unused materializations from the graph. |
-void AllocationSinking::RemoveUnusedMaterializations() { |
- intptr_t j = 0; |
- for (intptr_t i = 0; i < materializations_.length(); i++) { |
- MaterializeObjectInstr* mat = materializations_[i]; |
- if ((mat->input_use_list() == NULL) && (mat->env_use_list() == NULL)) { |
- // Check if this materialization failed to compute and remove any |
- // unforwarded loads. There were no loads from any allocation sinking |
- // candidate in the beggining so it is safe to assume that any encountered |
- // load was inserted by CreateMaterializationAt. |
- for (intptr_t i = 0; i < mat->InputCount(); i++) { |
- LoadFieldInstr* load = mat->InputAt(i)->definition()->AsLoadField(); |
- if ((load != NULL) && |
- (load->instance()->definition() == mat->allocation())) { |
- load->ReplaceUsesWith(flow_graph_->constant_null()); |
- load->RemoveFromGraph(); |
- } |
- } |
- mat->RemoveFromGraph(); |
- } else { |
- if (j != i) { |
- materializations_[j] = mat; |
- } |
- j++; |
- } |
- } |
- materializations_.TruncateTo(j); |
-} |
- |
- |
-// Some candidates might stop being eligible for allocation sinking after |
-// the load forwarding because they flow into phis that load forwarding |
-// inserts. Discover such allocations and remove them from the list |
-// of allocation sinking candidates undoing all changes that we did |
-// in preparation for sinking these allocations. |
-void AllocationSinking::DiscoverFailedCandidates() { |
- // Transitively unmark all candidates that are not strictly valid. |
- bool changed; |
- do { |
- changed = false; |
- for (intptr_t i = 0; i < candidates_.length(); i++) { |
- Definition* alloc = candidates_[i]; |
- if (alloc->Identity().IsAllocationSinkingCandidate()) { |
- if (!IsAllocationSinkingCandidate(alloc, kStrictCheck)) { |
- alloc->SetIdentity(AliasIdentity::Unknown()); |
- changed = true; |
- } |
- } |
- } |
- } while (changed); |
- |
- // Remove all failed candidates from the candidates list. |
- intptr_t j = 0; |
- for (intptr_t i = 0; i < candidates_.length(); i++) { |
- Definition* alloc = candidates_[i]; |
- if (!alloc->Identity().IsAllocationSinkingCandidate()) { |
- if (FLAG_trace_optimization) { |
- OS::Print("allocation v%" Pd " can't be eliminated\n", |
- alloc->ssa_temp_index()); |
- } |
- |
-#ifdef DEBUG |
- for (Value* use = alloc->env_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- ASSERT(use->instruction()->IsMaterializeObject()); |
- } |
-#endif |
- |
- // All materializations will be removed from the graph. Remove inserted |
- // loads first and detach materializations from allocation's environment |
- // use list: we will reconstruct it when we start removing |
- // materializations. |
- alloc->set_env_use_list(NULL); |
- for (Value* use = alloc->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- if (use->instruction()->IsLoadField()) { |
- LoadFieldInstr* load = use->instruction()->AsLoadField(); |
- load->ReplaceUsesWith(flow_graph_->constant_null()); |
- load->RemoveFromGraph(); |
- } else { |
- ASSERT(use->instruction()->IsMaterializeObject() || |
- use->instruction()->IsPhi() || |
- use->instruction()->IsStoreInstanceField()); |
- } |
- } |
- } else { |
- if (j != i) { |
- candidates_[j] = alloc; |
- } |
- j++; |
- } |
- } |
- |
- if (j != candidates_.length()) { // Something was removed from candidates. |
- intptr_t k = 0; |
- for (intptr_t i = 0; i < materializations_.length(); i++) { |
- MaterializeObjectInstr* mat = materializations_[i]; |
- if (!mat->allocation()->Identity().IsAllocationSinkingCandidate()) { |
- // Restore environment uses of the allocation that were replaced |
- // by this materialization and drop materialization. |
- mat->ReplaceUsesWith(mat->allocation()); |
- mat->RemoveFromGraph(); |
- } else { |
- if (k != i) { |
- materializations_[k] = mat; |
- } |
- k++; |
- } |
- } |
- materializations_.TruncateTo(k); |
- } |
- |
- candidates_.TruncateTo(j); |
-} |
- |
- |
-void AllocationSinking::Optimize() { |
- CollectCandidates(); |
- |
- // Insert MaterializeObject instructions that will describe the state of the |
- // object at all deoptimization points. Each inserted materialization looks |
- // like this (where v_0 is allocation that we are going to eliminate): |
- // v_1 <- LoadField(v_0, field_1) |
- // ... |
- // v_N <- LoadField(v_0, field_N) |
- // v_{N+1} <- MaterializeObject(field_1 = v_1, ..., field_N = v_{N}) |
- for (intptr_t i = 0; i < candidates_.length(); i++) { |
- InsertMaterializations(candidates_[i]); |
- } |
- |
- // Run load forwarding to eliminate LoadField instructions inserted above. |
- // All loads will be successfully eliminated because: |
- // a) they use fields (not offsets) and thus provide precise aliasing |
- // information |
- // b) candidate does not escape and thus its fields is not affected by |
- // external effects from calls. |
- LoadOptimizer::OptimizeGraph(flow_graph_); |
- |
- NormalizeMaterializations(); |
- |
- RemoveUnusedMaterializations(); |
- |
- // If any candidates are no longer eligible for allocation sinking abort |
- // the optimization for them and undo any changes we did in preparation. |
- DiscoverFailedCandidates(); |
- |
- // At this point we have computed the state of object at each deoptimization |
- // point and we can eliminate it. Loads inserted above were forwarded so there |
- // are no uses of the allocation just as in the begging of the pass. |
- for (intptr_t i = 0; i < candidates_.length(); i++) { |
- EliminateAllocation(candidates_[i]); |
- } |
- |
- // Process materializations and unbox their arguments: materializations |
- // are part of the environment and can materialize boxes for double/mint/simd |
- // values when needed. |
- // TODO(vegorov): handle all box types here. |
- for (intptr_t i = 0; i < materializations_.length(); i++) { |
- MaterializeObjectInstr* mat = materializations_[i]; |
- for (intptr_t j = 0; j < mat->InputCount(); j++) { |
- Definition* defn = mat->InputAt(j)->definition(); |
- if (defn->IsBox()) { |
- mat->InputAt(j)->BindTo(defn->InputAt(0)->definition()); |
- } |
- } |
- } |
-} |
- |
- |
-// Remove materializations from the graph. Register allocator will treat them |
-// as part of the environment not as a real instruction. |
-void AllocationSinking::DetachMaterializations() { |
- for (intptr_t i = 0; i < materializations_.length(); i++) { |
- materializations_[i]->previous()->LinkTo(materializations_[i]->next()); |
- } |
-} |
- |
- |
-// Add a field/offset to the list of fields if it is not yet present there. |
-static bool AddSlot(ZoneGrowableArray<const Object*>* slots, |
- const Object& slot) { |
- for (intptr_t i = 0; i < slots->length(); i++) { |
- if ((*slots)[i]->raw() == slot.raw()) { |
- return false; |
- } |
- } |
- slots->Add(&slot); |
- return true; |
-} |
- |
- |
-// Find deoptimization exit for the given materialization assuming that all |
-// materializations are emitted right before the instruction which is a |
-// deoptimization exit. |
-static Instruction* ExitForMaterialization(MaterializeObjectInstr* mat) { |
- while (mat->next()->IsMaterializeObject()) { |
- mat = mat->next()->AsMaterializeObject(); |
- } |
- return mat->next(); |
-} |
- |
- |
-// Given the deoptimization exit find first materialization that was inserted |
-// before it. |
-static Instruction* FirstMaterializationAt(Instruction* exit) { |
- while (exit->previous()->IsMaterializeObject()) { |
- exit = exit->previous(); |
- } |
- return exit; |
-} |
- |
- |
-// Given the allocation and deoptimization exit try to find MaterializeObject |
-// instruction corresponding to this allocation at this exit. |
-MaterializeObjectInstr* AllocationSinking::MaterializationFor( |
- Definition* alloc, Instruction* exit) { |
- if (exit->IsMaterializeObject()) { |
- exit = ExitForMaterialization(exit->AsMaterializeObject()); |
- } |
- |
- for (MaterializeObjectInstr* mat = exit->previous()->AsMaterializeObject(); |
- mat != NULL; |
- mat = mat->previous()->AsMaterializeObject()) { |
- if (mat->allocation() == alloc) { |
- return mat; |
- } |
- } |
- |
- return NULL; |
-} |
- |
- |
-// Insert MaterializeObject instruction for the given allocation before |
-// the given instruction that can deoptimize. |
-void AllocationSinking::CreateMaterializationAt( |
- Instruction* exit, |
- Definition* alloc, |
- const ZoneGrowableArray<const Object*>& slots) { |
- ZoneGrowableArray<Value*>* values = |
- new(I) ZoneGrowableArray<Value*>(slots.length()); |
- |
- // All loads should be inserted before the first materialization so that |
- // IR follows the following pattern: loads, materializations, deoptimizing |
- // instruction. |
- Instruction* load_point = FirstMaterializationAt(exit); |
- |
- // Insert load instruction for every field. |
- for (intptr_t i = 0; i < slots.length(); i++) { |
- LoadFieldInstr* load = slots[i]->IsField() |
- ? new(I) LoadFieldInstr( |
- new(I) Value(alloc), |
- &Field::Cast(*slots[i]), |
- AbstractType::ZoneHandle(I), |
- alloc->token_pos()) |
- : new(I) LoadFieldInstr( |
- new(I) Value(alloc), |
- Smi::Cast(*slots[i]).Value(), |
- AbstractType::ZoneHandle(I), |
- alloc->token_pos()); |
- flow_graph_->InsertBefore( |
- load_point, load, NULL, FlowGraph::kValue); |
- values->Add(new(I) Value(load)); |
- } |
- |
- MaterializeObjectInstr* mat = NULL; |
- if (alloc->IsAllocateObject()) { |
- mat = new(I) MaterializeObjectInstr( |
- alloc->AsAllocateObject(), slots, values); |
- } else { |
- ASSERT(alloc->IsAllocateUninitializedContext()); |
- mat = new(I) MaterializeObjectInstr( |
- alloc->AsAllocateUninitializedContext(), slots, values); |
- } |
- |
- flow_graph_->InsertBefore(exit, mat, NULL, FlowGraph::kValue); |
- |
- // Replace all mentions of this allocation with a newly inserted |
- // MaterializeObject instruction. |
- // We must preserve the identity: all mentions are replaced by the same |
- // materialization. |
- for (Environment::DeepIterator env_it(exit->env()); |
- !env_it.Done(); |
- env_it.Advance()) { |
- Value* use = env_it.CurrentValue(); |
- if (use->definition() == alloc) { |
- use->RemoveFromUseList(); |
- use->set_definition(mat); |
- mat->AddEnvUse(use); |
- } |
- } |
- |
- // Mark MaterializeObject as an environment use of this allocation. |
- // This will allow us to discover it when we are looking for deoptimization |
- // exits for another allocation that potentially flows into this one. |
- Value* val = new(I) Value(alloc); |
- val->set_instruction(mat); |
- alloc->AddEnvUse(val); |
- |
- // Record inserted materialization. |
- materializations_.Add(mat); |
-} |
- |
- |
-// Add given instruction to the list of the instructions if it is not yet |
-// present there. |
-template<typename T> |
-void AddInstruction(GrowableArray<T*>* list, T* value) { |
- ASSERT(!value->IsGraphEntry()); |
- for (intptr_t i = 0; i < list->length(); i++) { |
- if ((*list)[i] == value) { |
- return; |
- } |
- } |
- list->Add(value); |
-} |
- |
- |
-// Transitively collect all deoptimization exits that might need this allocation |
-// rematerialized. It is not enough to collect only environment uses of this |
-// allocation because it can flow into other objects that will be |
-// dematerialized and that are referenced by deopt environments that |
-// don't contain this allocation explicitly. |
-void AllocationSinking::ExitsCollector::Collect(Definition* alloc) { |
- for (Value* use = alloc->env_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- if (use->instruction()->IsMaterializeObject()) { |
- AddInstruction(&exits_, ExitForMaterialization( |
- use->instruction()->AsMaterializeObject())); |
- } else { |
- AddInstruction(&exits_, use->instruction()); |
- } |
- } |
- |
- // Check if this allocation is stored into any other allocation sinking |
- // candidate and put it on worklist so that we conservatively collect all |
- // exits for that candidate as well because they potentially might see |
- // this object. |
- for (Value* use = alloc->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- Definition* obj = StoreInto(use); |
- if ((obj != NULL) && (obj != alloc)) { |
- AddInstruction(&worklist_, obj); |
- } |
- } |
-} |
- |
- |
-void AllocationSinking::ExitsCollector::CollectTransitively(Definition* alloc) { |
- exits_.TruncateTo(0); |
- worklist_.TruncateTo(0); |
- |
- worklist_.Add(alloc); |
- |
- // Note: worklist potentially will grow while we are iterating over it. |
- // We are not removing allocations from the worklist not to waste space on |
- // the side maintaining BitVector of already processed allocations: worklist |
- // is expected to be very small thus linear search in it is just as effecient |
- // as a bitvector. |
- for (intptr_t i = 0; i < worklist_.length(); i++) { |
- Collect(worklist_[i]); |
- } |
-} |
- |
- |
-void AllocationSinking::InsertMaterializations(Definition* alloc) { |
- // Collect all fields that are written for this instance. |
- ZoneGrowableArray<const Object*>* slots = |
- new(I) ZoneGrowableArray<const Object*>(5); |
- |
- for (Value* use = alloc->input_use_list(); |
- use != NULL; |
- use = use->next_use()) { |
- StoreInstanceFieldInstr* store = use->instruction()->AsStoreInstanceField(); |
- if ((store != NULL) && (store->instance()->definition() == alloc)) { |
- if (!store->field().IsNull()) { |
- AddSlot(slots, store->field()); |
- } else { |
- AddSlot(slots, Smi::ZoneHandle(I, Smi::New(store->offset_in_bytes()))); |
- } |
- } |
- } |
- |
- if (alloc->ArgumentCount() > 0) { |
- AllocateObjectInstr* alloc_object = alloc->AsAllocateObject(); |
- ASSERT(alloc_object->ArgumentCount() == 1); |
- intptr_t type_args_offset = |
- alloc_object->cls().type_arguments_field_offset(); |
- AddSlot(slots, Smi::ZoneHandle(I, Smi::New(type_args_offset))); |
- } |
- |
- // Collect all instructions that mention this object in the environment. |
- exits_collector_.CollectTransitively(alloc); |
- |
- // Insert materializations at environment uses. |
- for (intptr_t i = 0; i < exits_collector_.exits().length(); i++) { |
- CreateMaterializationAt( |
- exits_collector_.exits()[i], alloc, *slots); |
- } |
-} |
- |
- |
} // namespace dart |