Index: runtime/vm/intermediate_language_arm.cc |
diff --git a/runtime/vm/intermediate_language_arm.cc b/runtime/vm/intermediate_language_arm.cc |
index 8218494cdf44e5b926f83dd33373a976259238e8..977763a9f7b2749f376a2cd5581cc834be60b806 100644 |
--- a/runtime/vm/intermediate_language_arm.cc |
+++ b/runtime/vm/intermediate_language_arm.cc |
@@ -1583,16 +1583,16 @@ void GuardFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
class StoreInstanceFieldSlowPath : public SlowPathCode { |
public: |
- explicit StoreInstanceFieldSlowPath(StoreInstanceFieldInstr* instruction) |
- : instruction_(instruction) { } |
+ StoreInstanceFieldSlowPath(StoreInstanceFieldInstr* instruction, |
+ const Class& cls) |
+ : instruction_(instruction), cls_(cls) { } |
virtual void EmitNativeCode(FlowGraphCompiler* compiler) { |
__ Comment("StoreInstanceFieldSlowPath"); |
__ Bind(entry_label()); |
- const Class& double_class = compiler->double_class(); |
const Code& stub = |
- Code::Handle(StubCode::GetAllocationStubForClass(double_class)); |
- const ExternalLabel label(double_class.ToCString(), stub.EntryPoint()); |
+ Code::Handle(StubCode::GetAllocationStubForClass(cls_)); |
+ const ExternalLabel label(cls_.ToCString(), stub.EntryPoint()); |
LocationSummary* locs = instruction_->locs(); |
locs->live_registers()->Remove(locs->out()); |
@@ -1610,6 +1610,7 @@ class StoreInstanceFieldSlowPath : public SlowPathCode { |
private: |
StoreInstanceFieldInstr* instruction_; |
+ const Class& cls_; |
}; |
@@ -1653,12 +1654,22 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
DRegister value = EvenDRegisterOf(locs()->in(1).fpu_reg()); |
Register temp = locs()->temp(0).reg(); |
Register temp2 = locs()->temp(1).reg(); |
+ const intptr_t cid = field().UnboxedFieldCid(); |
if (is_initialization_) { |
+ const Class* cls = NULL; |
+ switch (cid) { |
+ case kDoubleCid: |
+ cls = &compiler->double_class(); |
+ break; |
+ // TODO(johnmccutchan): Add kFloat32x4Cid here. |
+ default: |
+ UNREACHABLE(); |
+ } |
StoreInstanceFieldSlowPath* slow_path = |
- new StoreInstanceFieldSlowPath(this); |
+ new StoreInstanceFieldSlowPath(this, *cls); |
compiler->AddSlowPathCode(slow_path); |
- __ TryAllocate(compiler->double_class(), |
+ __ TryAllocate(*cls, |
slow_path->entry_label(), |
temp, |
temp2); |
@@ -1670,7 +1681,15 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
} else { |
__ ldr(temp, FieldAddress(instance_reg, field().Offset())); |
} |
- __ StoreDToOffset(value, temp, Double::value_offset() - kHeapObjectTag); |
+ switch (cid) { |
+ case kDoubleCid: |
+ __ StoreDToOffset(value, temp, Double::value_offset() - kHeapObjectTag); |
+ // TODO(johnmccutchan): Add kFloat32x4Cid here. |
+ break; |
+ default: |
+ UNREACHABLE(); |
+ } |
+ |
return; |
} |
@@ -1680,25 +1699,36 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
Register temp2 = locs()->temp(1).reg(); |
DRegister fpu_temp = EvenDRegisterOf(locs()->temp(2).fpu_reg()); |
- Label store_pointer, copy_payload; |
+ Label store_pointer; |
+ Label copy_double; |
+ Label store_double; |
+ |
__ LoadObject(temp, Field::ZoneHandle(field().raw())); |
- __ ldr(temp2, FieldAddress(temp, Field::guarded_cid_offset())); |
- __ CompareImmediate(temp2, kDoubleCid); |
- __ b(&store_pointer, NE); |
+ |
__ ldr(temp2, FieldAddress(temp, Field::is_nullable_offset())); |
__ CompareImmediate(temp2, kNullCid); |
__ b(&store_pointer, EQ); |
+ |
__ ldrb(temp2, FieldAddress(temp, Field::kind_bits_offset())); |
__ tst(temp2, ShifterOperand(1 << Field::kUnboxingCandidateBit)); |
__ b(&store_pointer, EQ); |
+ __ ldr(temp2, FieldAddress(temp, Field::guarded_cid_offset())); |
+ __ CompareImmediate(temp2, kDoubleCid); |
+ __ b(&store_double, EQ); |
+ |
+ // Fall through. |
+ __ b(&store_pointer); |
+ |
+ __ Bind(&store_double); |
+ |
__ ldr(temp, FieldAddress(instance_reg, field().Offset())); |
__ CompareImmediate(temp, |
reinterpret_cast<intptr_t>(Object::null())); |
- __ b(©_payload, NE); |
+ __ b(©_double, NE); |
StoreInstanceFieldSlowPath* slow_path = |
- new StoreInstanceFieldSlowPath(this); |
+ new StoreInstanceFieldSlowPath(this, compiler->double_class()); |
compiler->AddSlowPathCode(slow_path); |
if (!compiler->is_optimizing()) { |
@@ -1715,7 +1745,7 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
__ StoreIntoObject(instance_reg, |
FieldAddress(instance_reg, field().Offset()), |
temp2); |
- __ Bind(©_payload); |
+ __ Bind(©_double); |
__ LoadDFromOffset(fpu_temp, |
value_reg, |
Double::value_offset() - kHeapObjectTag); |
@@ -1926,7 +1956,16 @@ void LoadFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
DRegister result = EvenDRegisterOf(locs()->out().fpu_reg()); |
Register temp = locs()->temp(0).reg(); |
__ ldr(temp, FieldAddress(instance_reg, offset_in_bytes())); |
- __ LoadDFromOffset(result, temp, Double::value_offset() - kHeapObjectTag); |
+ intptr_t cid = field()->UnboxedFieldCid(); |
+ switch (cid) { |
+ case kDoubleCid: |
+ __ LoadDFromOffset(result, temp, |
+ Double::value_offset() - kHeapObjectTag); |
+ break; |
+ // TODO(johnmccutchan): Add Float32x4 path here. |
+ default: |
+ UNREACHABLE(); |
+ } |
return; |
} |
@@ -1937,20 +1976,27 @@ void LoadFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
DRegister value = EvenDRegisterOf(locs()->temp(0).fpu_reg()); |
Label load_pointer; |
+ Label load_double; |
+ |
__ LoadObject(result_reg, Field::ZoneHandle(field()->raw())); |
FieldAddress field_cid_operand(result_reg, Field::guarded_cid_offset()); |
FieldAddress field_nullability_operand(result_reg, |
Field::is_nullable_offset()); |
- __ ldr(temp, field_cid_operand); |
- __ CompareImmediate(temp, kDoubleCid); |
- __ b(&load_pointer, NE); |
- |
__ ldr(temp, field_nullability_operand); |
__ CompareImmediate(temp, kNullCid); |
__ b(&load_pointer, EQ); |
+ __ ldr(temp, field_cid_operand); |
+ __ CompareImmediate(temp, kDoubleCid); |
+ __ b(&load_double, EQ); |
+ |
+ // Fall through. |
+ __ b(&load_pointer); |
+ |
+ __ Bind(&load_double); |
+ |
BoxDoubleSlowPath* slow_path = new BoxDoubleSlowPath(this); |
compiler->AddSlowPathCode(slow_path); |
@@ -1969,6 +2015,9 @@ void LoadFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
result_reg, |
Double::value_offset() - kHeapObjectTag); |
__ b(&done); |
+ |
+ // TODO(johnmccutchan): Add Float32x4 path here. |
+ |
__ Bind(&load_pointer); |
} |
__ LoadFromOffset(kWord, result_reg, |