| Index: runtime/vm/flow_graph_compiler_arm.cc
 | 
| diff --git a/runtime/vm/flow_graph_compiler_arm.cc b/runtime/vm/flow_graph_compiler_arm.cc
 | 
| index 7347085a29e1fbd5ea51567a78ff158c1991cb06..1bb28294528eff35bd122d7a013c7748ff16396a 100644
 | 
| --- a/runtime/vm/flow_graph_compiler_arm.cc
 | 
| +++ b/runtime/vm/flow_graph_compiler_arm.cc
 | 
| @@ -1600,7 +1600,11 @@ void ParallelMoveResolver::EmitMove(int index) {
 | 
|      ASSERT(source.IsConstant());
 | 
|      const Object& constant = source.constant();
 | 
|      if (destination.IsRegister()) {
 | 
| -      __ LoadObject(destination.reg(), constant);
 | 
| +      if (source.constant_instruction()->representation() == kUnboxedInt32) {
 | 
| +        __ LoadImmediate(destination.reg(), Smi::Cast(constant).Value());
 | 
| +      } else {
 | 
| +        __ LoadObject(destination.reg(), constant);
 | 
| +      }
 | 
|      } else if (destination.IsFpuRegister()) {
 | 
|        const DRegister dst = EvenDRegisterOf(destination.fpu_reg());
 | 
|        if (Utils::DoublesBitEqual(Double::Cast(constant).value(), 0.0) &&
 | 
| @@ -1626,7 +1630,11 @@ void ParallelMoveResolver::EmitMove(int index) {
 | 
|      } else {
 | 
|        ASSERT(destination.IsStackSlot());
 | 
|        const intptr_t dest_offset = destination.ToStackSlotOffset();
 | 
| -      __ LoadObject(TMP, constant);
 | 
| +      if (source.constant_instruction()->representation() == kUnboxedInt32) {
 | 
| +        __ LoadImmediate(TMP, Smi::Cast(constant).Value());
 | 
| +      } else {
 | 
| +        __ LoadObject(TMP, constant);
 | 
| +      }
 | 
|        __ StoreToOffset(kWord, TMP, FP, dest_offset);
 | 
|      }
 | 
|    }
 | 
| 
 |