| Index: runtime/vm/intermediate_language_ia32.cc
|
| diff --git a/runtime/vm/intermediate_language_ia32.cc b/runtime/vm/intermediate_language_ia32.cc
|
| index c1b3e0435bfc0b34ac238c4fff8c7fa10052858e..5655f75ee9626477a74f01ed5bf04e88e8b8ccc3 100644
|
| --- a/runtime/vm/intermediate_language_ia32.cc
|
| +++ b/runtime/vm/intermediate_language_ia32.cc
|
| @@ -1590,36 +1590,74 @@ void GuardFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| }
|
|
|
|
|
| +bool Field::IsUnboxedField() const {
|
| + bool valid_class = (guarded_cid() == kDoubleCid) ||
|
| + (guarded_cid() == kFloat32x4Cid);
|
| + return is_unboxing_candidate() && !is_final() && !is_nullable() &&
|
| + valid_class;
|
| +}
|
| +
|
| +
|
| class StoreInstanceFieldSlowPath : public SlowPathCode {
|
| public:
|
| - StoreInstanceFieldSlowPath(StoreInstanceFieldInstr* instruction,
|
| - const Class& cls)
|
| - : instruction_(instruction), cls_(cls) { }
|
| + explicit StoreInstanceFieldSlowPath(StoreInstanceFieldInstr* instruction)
|
| + : instruction_(instruction) { }
|
|
|
| virtual void EmitNativeCode(FlowGraphCompiler* compiler) {
|
| __ Comment("StoreInstanceFieldSlowPath");
|
| - __ Bind(entry_label());
|
| - const Code& stub =
|
| - Code::Handle(StubCode::GetAllocationStubForClass(cls_));
|
| - const ExternalLabel label(cls_.ToCString(), stub.EntryPoint());
|
| -
|
| - LocationSummary* locs = instruction_->locs();
|
| - locs->live_registers()->Remove(locs->out());
|
| -
|
| - compiler->SaveLiveRegisters(locs);
|
| - compiler->GenerateCall(Scanner::kNoSourcePos, // No token position.
|
| - &label,
|
| - PcDescriptors::kOther,
|
| - locs);
|
| - __ MoveRegister(locs->temp(0).reg(), EAX);
|
| - compiler->RestoreLiveRegisters(locs);
|
| + {
|
| + __ Bind(double_entry_label());
|
| + const Class& cls = compiler->double_class();
|
| + const Code& stub =
|
| + Code::Handle(StubCode::GetAllocationStubForClass(cls));
|
| + const ExternalLabel label(cls.ToCString(), stub.EntryPoint());
|
| +
|
| + LocationSummary* locs = instruction_->locs();
|
| + locs->live_registers()->Remove(locs->out());
|
| + compiler->SaveLiveRegisters(locs);
|
| + compiler->GenerateCall(Scanner::kNoSourcePos, // No token position.
|
| + &label,
|
| + PcDescriptors::kOther,
|
| + locs);
|
| + __ MoveRegister(locs->temp(0).reg(), EAX);
|
| + compiler->RestoreLiveRegisters(locs);
|
| + __ jmp(double_exit_label());
|
| + }
|
| + {
|
| + __ Bind(float32x4_entry_label());
|
| + const Class& cls = compiler->float32x4_class();
|
| + const Code& stub =
|
| + Code::Handle(StubCode::GetAllocationStubForClass(cls));
|
| + const ExternalLabel label(cls.ToCString(), stub.EntryPoint());
|
| + LocationSummary* locs = instruction_->locs();
|
| + locs->live_registers()->Remove(locs->out());
|
| + compiler->SaveLiveRegisters(locs);
|
| + compiler->GenerateCall(Scanner::kNoSourcePos, // No token position.
|
| + &label,
|
| + PcDescriptors::kOther,
|
| + locs);
|
| + __ MoveRegister(locs->temp(0).reg(), EAX);
|
| + compiler->RestoreLiveRegisters(locs);
|
| + __ jmp(float32x4_exit_label());
|
| + }
|
| + }
|
|
|
| - __ jmp(exit_label());
|
| + Label* double_entry_label() {
|
| + // Use default SlowPathCode label for double.
|
| + return entry_label();
|
| }
|
| + Label* double_exit_label() {
|
| + // Use default SlowPathCode label for double.
|
| + return exit_label();
|
| + }
|
| +
|
| + Label* float32x4_entry_label() { return &float32x4_entry_label_; }
|
| + Label* float32x4_exit_label() { return &float32x4_exit_label_; }
|
|
|
| private:
|
| + Label float32x4_entry_label_;
|
| + Label float32x4_exit_label_;
|
| StoreInstanceFieldInstr* instruction_;
|
| - const Class& cls_;
|
| };
|
|
|
|
|
| @@ -1666,25 +1704,34 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| const intptr_t cid = field().UnboxedFieldCid();
|
|
|
| if (is_initialization_) {
|
| + StoreInstanceFieldSlowPath* slow_path =
|
| + new StoreInstanceFieldSlowPath(this);
|
| + compiler->AddSlowPathCode(slow_path);
|
| +
|
| const Class* cls = NULL;
|
| + Label* entry_label = NULL;
|
| + Label* exit_label = NULL;
|
| switch (cid) {
|
| case kDoubleCid:
|
| cls = &compiler->double_class();
|
| + entry_label = slow_path->double_entry_label();
|
| + exit_label = slow_path->double_exit_label();
|
| + break;
|
| + case kFloat32x4Cid:
|
| + cls = &compiler->float32x4_class();
|
| + entry_label = slow_path->float32x4_entry_label();
|
| + exit_label = slow_path->float32x4_exit_label();
|
| break;
|
| - // TODO(johnmccutchan): Add kFloat32x4Cid here.
|
| default:
|
| UNREACHABLE();
|
| }
|
| - StoreInstanceFieldSlowPath* slow_path =
|
| - new StoreInstanceFieldSlowPath(this, *cls);
|
| - compiler->AddSlowPathCode(slow_path);
|
|
|
| __ TryAllocate(*cls,
|
| - slow_path->entry_label(),
|
| + entry_label,
|
| Assembler::kFarJump,
|
| temp,
|
| temp2);
|
| - __ Bind(slow_path->exit_label());
|
| + __ Bind(exit_label);
|
| __ movl(temp2, temp);
|
| __ StoreIntoObject(instance_reg,
|
| FieldAddress(instance_reg, field().Offset()),
|
| @@ -1694,8 +1741,12 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| }
|
| switch (cid) {
|
| case kDoubleCid:
|
| - __ movsd(FieldAddress(temp, Double::value_offset()), value);
|
| - // TODO(johnmccutchan): Add kFloat32x4Cid here.
|
| + __ Comment("UnboxedDoubleStoreInstanceFieldInstr");
|
| + __ movsd(FieldAddress(temp, Double::value_offset()), value);
|
| + break;
|
| + case kFloat32x4Cid:
|
| + __ Comment("UnboxedFloat32x4StoreInstanceFieldInstr");
|
| + __ movups(FieldAddress(temp, Float32x4::value_offset()), value);
|
| break;
|
| default:
|
| UNREACHABLE();
|
| @@ -1710,8 +1761,8 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| FpuRegister fpu_temp = locs()->temp(2).fpu_reg();
|
|
|
| Label store_pointer;
|
| - Label copy_double;
|
| Label store_double;
|
| + Label store_float32x4;
|
|
|
| __ LoadObject(temp, Field::ZoneHandle(field().raw()));
|
|
|
| @@ -1727,41 +1778,80 @@ void StoreInstanceFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| Immediate(kDoubleCid));
|
| __ j(EQUAL, &store_double);
|
|
|
| + __ cmpl(FieldAddress(temp, Field::guarded_cid_offset()),
|
| + Immediate(kFloat32x4Cid));
|
| + __ j(EQUAL, &store_float32x4);
|
| +
|
| // Fall through.
|
| __ jmp(&store_pointer);
|
|
|
| - __ Bind(&store_double);
|
| -
|
| - const Immediate& raw_null =
|
| - Immediate(reinterpret_cast<intptr_t>(Object::null()));
|
| - __ movl(temp, FieldAddress(instance_reg, field().Offset()));
|
| - __ cmpl(temp, raw_null);
|
| - __ j(NOT_EQUAL, ©_double);
|
| -
|
| StoreInstanceFieldSlowPath* slow_path =
|
| - new StoreInstanceFieldSlowPath(this, compiler->double_class());
|
| + new StoreInstanceFieldSlowPath(this);
|
| compiler->AddSlowPathCode(slow_path);
|
|
|
| - if (!compiler->is_optimizing()) {
|
| - locs()->live_registers()->Add(locs()->in(0));
|
| - locs()->live_registers()->Add(locs()->in(1));
|
| + {
|
| + __ Bind(&store_double);
|
| + Label copy_double;
|
| +
|
| + const Immediate& raw_null =
|
| + Immediate(reinterpret_cast<intptr_t>(Object::null()));
|
| + __ movl(temp, FieldAddress(instance_reg, field().Offset()));
|
| + __ cmpl(temp, raw_null);
|
| + __ j(NOT_EQUAL, ©_double);
|
| +
|
| + if (!compiler->is_optimizing()) {
|
| + locs()->live_registers()->Add(locs()->in(0));
|
| + locs()->live_registers()->Add(locs()->in(1));
|
| + }
|
| +
|
| + __ TryAllocate(compiler->double_class(),
|
| + slow_path->double_entry_label(),
|
| + Assembler::kFarJump,
|
| + temp,
|
| + temp2);
|
| + __ Bind(slow_path->double_exit_label());
|
| + __ movl(temp2, temp);
|
| + __ StoreIntoObject(instance_reg,
|
| + FieldAddress(instance_reg, field().Offset()),
|
| + temp2);
|
| +
|
| + __ Bind(©_double);
|
| + __ movsd(fpu_temp, FieldAddress(value_reg, Double::value_offset()));
|
| + __ movsd(FieldAddress(temp, Double::value_offset()), fpu_temp);
|
| + __ jmp(&skip_store);
|
| }
|
|
|
| - __ TryAllocate(compiler->double_class(),
|
| - slow_path->entry_label(),
|
| - Assembler::kFarJump,
|
| - temp,
|
| - temp2);
|
| - __ Bind(slow_path->exit_label());
|
| - __ movl(temp2, temp);
|
| - __ StoreIntoObject(instance_reg,
|
| - FieldAddress(instance_reg, field().Offset()),
|
| - temp2);
|
| + {
|
| + __ Bind(&store_float32x4);
|
| + Label copy_float32x4;
|
| +
|
| + const Immediate& raw_null =
|
| + Immediate(reinterpret_cast<intptr_t>(Object::null()));
|
| + __ movl(temp, FieldAddress(instance_reg, field().Offset()));
|
| + __ cmpl(temp, raw_null);
|
| + __ j(NOT_EQUAL, ©_float32x4);
|
| +
|
| + if (!compiler->is_optimizing()) {
|
| + locs()->live_registers()->Add(locs()->in(0));
|
| + locs()->live_registers()->Add(locs()->in(1));
|
| + }
|
| +
|
| + __ TryAllocate(compiler->float32x4_class(),
|
| + slow_path->float32x4_entry_label(),
|
| + Assembler::kFarJump,
|
| + temp,
|
| + temp2);
|
| + __ Bind(slow_path->float32x4_exit_label());
|
| + __ movl(temp2, temp);
|
| + __ StoreIntoObject(instance_reg,
|
| + FieldAddress(instance_reg, field().Offset()),
|
| + temp2);
|
|
|
| - __ Bind(©_double);
|
| - __ movsd(fpu_temp, FieldAddress(value_reg, Double::value_offset()));
|
| - __ movsd(FieldAddress(temp, Double::value_offset()), fpu_temp);
|
| - __ jmp(&skip_store);
|
| + __ Bind(©_float32x4);
|
| + __ movups(fpu_temp, FieldAddress(value_reg, Float32x4::value_offset()));
|
| + __ movups(FieldAddress(temp, Float32x4::value_offset()), fpu_temp);
|
| + __ jmp(&skip_store);
|
| + }
|
| __ Bind(&store_pointer);
|
| }
|
|
|
| @@ -1906,34 +1996,70 @@ void AllocateObjectWithBoundsCheckInstr::EmitNativeCode(
|
| }
|
|
|
|
|
| -class BoxDoubleSlowPath : public SlowPathCode {
|
| +class LoadFieldSlowPath : public SlowPathCode {
|
| public:
|
| - explicit BoxDoubleSlowPath(Instruction* instruction)
|
| + explicit LoadFieldSlowPath(Instruction* instruction)
|
| : instruction_(instruction) { }
|
|
|
| virtual void EmitNativeCode(FlowGraphCompiler* compiler) {
|
| - __ Comment("BoxDoubleSlowPath");
|
| - __ Bind(entry_label());
|
| - const Class& double_class = compiler->double_class();
|
| - const Code& stub =
|
| - Code::Handle(StubCode::GetAllocationStubForClass(double_class));
|
| - const ExternalLabel label(double_class.ToCString(), stub.EntryPoint());
|
| -
|
| - LocationSummary* locs = instruction_->locs();
|
| - locs->live_registers()->Remove(locs->out());
|
| -
|
| - compiler->SaveLiveRegisters(locs);
|
| - compiler->GenerateCall(Scanner::kNoSourcePos, // No token position.
|
| - &label,
|
| - PcDescriptors::kOther,
|
| - locs);
|
| - __ MoveRegister(locs->out().reg(), EAX);
|
| - compiler->RestoreLiveRegisters(locs);
|
| + __ Comment("LoadFieldSlowPath");
|
| + {
|
| + __ Bind(double_entry_label());
|
| + const Class& double_class = compiler->double_class();
|
| + const Code& stub =
|
| + Code::Handle(StubCode::GetAllocationStubForClass(double_class));
|
| + const ExternalLabel label(double_class.ToCString(), stub.EntryPoint());
|
| +
|
| + LocationSummary* locs = instruction_->locs();
|
| + locs->live_registers()->Remove(locs->out());
|
| +
|
| + compiler->SaveLiveRegisters(locs);
|
| + compiler->GenerateCall(Scanner::kNoSourcePos, // No token position.
|
| + &label,
|
| + PcDescriptors::kOther,
|
| + locs);
|
| + __ MoveRegister(locs->out().reg(), EAX);
|
| + compiler->RestoreLiveRegisters(locs);
|
| +
|
| + __ jmp(double_exit_label());
|
| + }
|
| + {
|
| + __ Bind(float32x4_entry_label());
|
| + const Class& float32x4_class = compiler->float32x4_class();
|
| + const Code& stub =
|
| + Code::Handle(StubCode::GetAllocationStubForClass(float32x4_class));
|
| + const ExternalLabel label(float32x4_class.ToCString(), stub.EntryPoint());
|
| +
|
| + LocationSummary* locs = instruction_->locs();
|
| + locs->live_registers()->Remove(locs->out());
|
| +
|
| + compiler->SaveLiveRegisters(locs);
|
| + compiler->GenerateCall(Scanner::kNoSourcePos, // No token position.
|
| + &label,
|
| + PcDescriptors::kOther,
|
| + locs);
|
| + __ MoveRegister(locs->out().reg(), EAX);
|
| + compiler->RestoreLiveRegisters(locs);
|
| +
|
| + __ jmp(float32x4_exit_label());
|
| + }
|
| + }
|
|
|
| - __ jmp(exit_label());
|
| + Label* double_entry_label() {
|
| + // Use default SlowPathCode label for double.
|
| + return entry_label();
|
| }
|
| + Label* double_exit_label() {
|
| + // Use default SlowPathCode label for double.
|
| + return exit_label();
|
| + }
|
| +
|
| + Label* float32x4_entry_label() { return &float32x4_entry_label_; }
|
| + Label* float32x4_exit_label() { return &float32x4_exit_label_; }
|
|
|
| private:
|
| + Label float32x4_entry_label_;
|
| + Label float32x4_exit_label_;
|
| Instruction* instruction_;
|
| };
|
|
|
| @@ -1971,9 +2097,13 @@ void LoadFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| const intptr_t cid = field()->UnboxedFieldCid();
|
| switch (cid) {
|
| case kDoubleCid:
|
| + __ Comment("UnboxedDoubleLoadFieldInstr");
|
| __ movsd(result, FieldAddress(temp, Double::value_offset()));
|
| break;
|
| - // TODO(johnmccutchan): Add Float32x4 path here.
|
| + case kFloat32x4Cid:
|
| + __ Comment("UnboxedFloat32x4LoadFieldInstr");
|
| + __ movups(result, FieldAddress(temp, Float32x4::value_offset()));
|
| + break;
|
| default:
|
| UNREACHABLE();
|
| }
|
| @@ -1985,9 +2115,13 @@ void LoadFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| if (IsPotentialUnboxedLoad()) {
|
| Register temp = locs()->temp(1).reg();
|
| XmmRegister value = locs()->temp(0).fpu_reg();
|
| + LoadFieldSlowPath* slow_path = new LoadFieldSlowPath(this);
|
| + compiler->AddSlowPathCode(slow_path);
|
|
|
| Label load_pointer;
|
| Label load_double;
|
| + Label load_float32x4;
|
| +
|
| __ LoadObject(result, Field::ZoneHandle(field()->raw()));
|
|
|
| FieldAddress field_cid_operand(result, Field::guarded_cid_offset());
|
| @@ -1999,29 +2133,49 @@ void LoadFieldInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| __ cmpl(field_cid_operand, Immediate(kDoubleCid));
|
| __ j(EQUAL, &load_double);
|
|
|
| + __ cmpl(field_cid_operand, Immediate(kFloat32x4Cid));
|
| + __ j(EQUAL, &load_float32x4);
|
| +
|
| // Fall through.
|
| __ jmp(&load_pointer);
|
|
|
| - __ Bind(&load_double);
|
| - BoxDoubleSlowPath* slow_path = new BoxDoubleSlowPath(this);
|
| - compiler->AddSlowPathCode(slow_path);
|
| + {
|
| + __ Bind(&load_double);
|
|
|
| - if (!compiler->is_optimizing()) {
|
| - locs()->live_registers()->Add(locs()->in(0));
|
| + if (!compiler->is_optimizing()) {
|
| + locs()->live_registers()->Add(locs()->in(0));
|
| + }
|
| +
|
| + __ TryAllocate(compiler->double_class(),
|
| + slow_path->double_entry_label(),
|
| + Assembler::kFarJump,
|
| + result,
|
| + temp);
|
| + __ Bind(slow_path->double_exit_label());
|
| + __ movl(temp, FieldAddress(instance_reg, offset_in_bytes()));
|
| + __ movsd(value, FieldAddress(temp, Double::value_offset()));
|
| + __ movsd(FieldAddress(result, Double::value_offset()), value);
|
| + __ jmp(&done);
|
| }
|
|
|
| - __ TryAllocate(compiler->double_class(),
|
| - slow_path->entry_label(),
|
| - Assembler::kFarJump,
|
| - result,
|
| - temp);
|
| - __ Bind(slow_path->exit_label());
|
| - __ movl(temp, FieldAddress(instance_reg, offset_in_bytes()));
|
| - __ movsd(value, FieldAddress(temp, Double::value_offset()));
|
| - __ movsd(FieldAddress(result, Double::value_offset()), value);
|
| - __ jmp(&done);
|
| + {
|
| + __ Bind(&load_float32x4);
|
| +
|
| + if (!compiler->is_optimizing()) {
|
| + locs()->live_registers()->Add(locs()->in(0));
|
| + }
|
|
|
| - // TODO(johnmccutchan): Add Float32x4 path here.
|
| + __ TryAllocate(compiler->float32x4_class(),
|
| + slow_path->float32x4_entry_label(),
|
| + Assembler::kFarJump,
|
| + result,
|
| + temp);
|
| + __ Bind(slow_path->float32x4_exit_label());
|
| + __ movl(temp, FieldAddress(instance_reg, offset_in_bytes()));
|
| + __ movups(value, FieldAddress(temp, Float32x4::value_offset()));
|
| + __ movups(FieldAddress(result, Float32x4::value_offset()), value);
|
| + __ jmp(&done);
|
| + }
|
|
|
| __ Bind(&load_pointer);
|
| }
|
| @@ -2877,6 +3031,38 @@ void CheckEitherNonSmiInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| }
|
|
|
|
|
| +class BoxDoubleSlowPath : public SlowPathCode {
|
| + public:
|
| + explicit BoxDoubleSlowPath(BoxDoubleInstr* instruction)
|
| + : instruction_(instruction) { }
|
| +
|
| + virtual void EmitNativeCode(FlowGraphCompiler* compiler) {
|
| + __ Comment("BoxDoubleSlowPath");
|
| + __ Bind(entry_label());
|
| + const Class& double_class = compiler->double_class();
|
| + const Code& stub =
|
| + Code::Handle(StubCode::GetAllocationStubForClass(double_class));
|
| + const ExternalLabel label(double_class.ToCString(), stub.EntryPoint());
|
| +
|
| + LocationSummary* locs = instruction_->locs();
|
| + locs->live_registers()->Remove(locs->out());
|
| +
|
| + compiler->SaveLiveRegisters(locs);
|
| + compiler->GenerateCall(Scanner::kNoSourcePos, // No token position.
|
| + &label,
|
| + PcDescriptors::kOther,
|
| + locs);
|
| + __ MoveRegister(locs->out().reg(), EAX);
|
| + compiler->RestoreLiveRegisters(locs);
|
| +
|
| + __ jmp(exit_label());
|
| + }
|
| +
|
| + private:
|
| + BoxDoubleInstr* instruction_;
|
| +};
|
| +
|
| +
|
| LocationSummary* BoxDoubleInstr::MakeLocationSummary(bool opt) const {
|
| const intptr_t kNumInputs = 1;
|
| const intptr_t kNumTemps = 0;
|
| @@ -2953,19 +3139,6 @@ void UnboxDoubleInstr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| }
|
|
|
|
|
| -LocationSummary* BoxFloat32x4Instr::MakeLocationSummary(bool opt) const {
|
| - const intptr_t kNumInputs = 1;
|
| - const intptr_t kNumTemps = 0;
|
| - LocationSummary* summary =
|
| - new LocationSummary(kNumInputs,
|
| - kNumTemps,
|
| - LocationSummary::kCallOnSlowPath);
|
| - summary->set_in(0, Location::RequiresFpuRegister());
|
| - summary->set_out(Location::RequiresRegister());
|
| - return summary;
|
| -}
|
| -
|
| -
|
| class BoxFloat32x4SlowPath : public SlowPathCode {
|
| public:
|
| explicit BoxFloat32x4SlowPath(BoxFloat32x4Instr* instruction)
|
| @@ -2998,6 +3171,19 @@ class BoxFloat32x4SlowPath : public SlowPathCode {
|
| };
|
|
|
|
|
| +LocationSummary* BoxFloat32x4Instr::MakeLocationSummary(bool opt) const {
|
| + const intptr_t kNumInputs = 1;
|
| + const intptr_t kNumTemps = 0;
|
| + LocationSummary* summary =
|
| + new LocationSummary(kNumInputs,
|
| + kNumTemps,
|
| + LocationSummary::kCallOnSlowPath);
|
| + summary->set_in(0, Location::RequiresFpuRegister());
|
| + summary->set_out(Location::RequiresRegister());
|
| + return summary;
|
| +}
|
| +
|
| +
|
| void BoxFloat32x4Instr::EmitNativeCode(FlowGraphCompiler* compiler) {
|
| BoxFloat32x4SlowPath* slow_path = new BoxFloat32x4SlowPath(this);
|
| compiler->AddSlowPathCode(slow_path);
|
|
|