Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(125)

Unified Diff: runtime/vm/flow_graph_compiler_mips.cc

Issue 365983002: Make isolate specific stub code accessors instance methods instead (Closed) Base URL: http://dart.googlecode.com/svn/branches/bleeding_edge/dart/
Patch Set: Created 6 years, 6 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « runtime/vm/flow_graph_compiler_ia32.cc ('k') | runtime/vm/flow_graph_compiler_x64.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: runtime/vm/flow_graph_compiler_mips.cc
===================================================================
--- runtime/vm/flow_graph_compiler_mips.cc (revision 37923)
+++ runtime/vm/flow_graph_compiler_mips.cc (working copy)
@@ -167,7 +167,8 @@
ASSERT(deopt_env() != NULL);
- __ BranchLink(&StubCode::DeoptimizeLabel());
+ StubCode* stub_code = compiler->isolate()->stub_code();
+ __ BranchLink(&stub_code->DeoptimizeLabel());
set_pc_offset(assem->CodeSize());
#undef __
}
@@ -204,18 +205,19 @@
ASSERT(temp_reg == kNoRegister); // Unused on MIPS.
const SubtypeTestCache& type_test_cache =
SubtypeTestCache::ZoneHandle(SubtypeTestCache::New());
+ StubCode* stub_code = isolate()->stub_code();
__ LoadObject(A2, type_test_cache);
if (test_kind == kTestTypeOneArg) {
ASSERT(type_arguments_reg == kNoRegister);
__ LoadImmediate(A1, reinterpret_cast<int32_t>(Object::null()));
- __ BranchLink(&StubCode::Subtype1TestCacheLabel());
+ __ BranchLink(&stub_code->Subtype1TestCacheLabel());
} else if (test_kind == kTestTypeTwoArgs) {
ASSERT(type_arguments_reg == kNoRegister);
__ LoadImmediate(A1, reinterpret_cast<int32_t>(Object::null()));
- __ BranchLink(&StubCode::Subtype2TestCacheLabel());
+ __ BranchLink(&stub_code->Subtype2TestCacheLabel());
} else if (test_kind == kTestTypeThreeArgs) {
ASSERT(type_arguments_reg == A1);
- __ BranchLink(&StubCode::Subtype3TestCacheLabel());
+ __ BranchLink(&stub_code->Subtype3TestCacheLabel());
} else {
UNREACHABLE();
}
@@ -914,13 +916,14 @@
__ Bind(&wrong_num_arguments);
if (function.IsClosureFunction()) {
// Invoke noSuchMethod function passing "call" as the original name.
+ StubCode* stub_code = isolate()->stub_code();
const int kNumArgsChecked = 1;
const ICData& ic_data = ICData::ZoneHandle(
ICData::New(function, Symbols::Call(), Object::empty_array(),
Isolate::kNoDeoptId, kNumArgsChecked));
__ LoadObject(S5, ic_data);
__ LeaveDartFrame(); // The arguments are still on the stack.
- __ Branch(&StubCode::CallNoSuchMethodFunctionLabel());
+ __ Branch(&stub_code->CallNoSuchMethodFunctionLabel());
// The noSuchMethod call may return to the caller, but not here.
__ break_(0);
} else if (check_correct_named_args) {
@@ -984,6 +987,7 @@
function.IsOptimizable() &&
(!is_optimizing() || may_reoptimize())) {
const Register function_reg = T0;
+ StubCode* stub_code = isolate()->stub_code();
__ GetNextPC(T2, TMP);
@@ -1020,7 +1024,7 @@
__ BranchSignedLess(T1, threshold, &dont_branch);
ASSERT(function_reg == T0);
- __ Branch(&StubCode::OptimizeFunctionLabel());
+ __ Branch(&stub_code->OptimizeFunctionLabel());
__ Bind(&dont_branch);
@@ -1060,6 +1064,7 @@
const int num_fixed_params = function.num_fixed_parameters();
const int num_copied_params = parsed_function().num_copied_params();
const int num_locals = parsed_function().num_stack_locals();
+ StubCode* stub_code = isolate()->stub_code();
// We check the number of passed arguments when we have to copy them due to
// the presence of optional parameters.
@@ -1099,7 +1104,7 @@
Isolate::kNoDeoptId, kNumArgsChecked));
__ LoadObject(S5, ic_data);
__ LeaveDartFrame(); // The arguments are still on the stack.
- __ Branch(&StubCode::CallNoSuchMethodFunctionLabel());
+ __ Branch(&stub_code->CallNoSuchMethodFunctionLabel());
// The noSuchMethod call may return to the caller, but not here.
__ break_(0);
} else {
@@ -1131,11 +1136,11 @@
// Emit function patching code. This will be swapped with the first 5 bytes
// at entry point.
patch_code_pc_offset_ = assembler()->CodeSize();
- __ BranchPatchable(&StubCode::FixCallersTargetLabel());
+ __ BranchPatchable(&stub_code->FixCallersTargetLabel());
if (is_optimizing()) {
lazy_deopt_pc_offset_ = assembler()->CodeSize();
- __ Branch(&StubCode::DeoptimizeLazyLabel());
+ __ Branch(&stub_code->DeoptimizeLazyLabel());
}
}
@@ -1333,10 +1338,11 @@
LocationSummary* locs,
const ICData& ic_data) {
uword label_address = 0;
+ StubCode* stub_code = isolate()->stub_code();
if (ic_data.NumArgsTested() == 0) {
- label_address = StubCode::ZeroArgsUnoptimizedStaticCallEntryPoint();
+ label_address = stub_code->ZeroArgsUnoptimizedStaticCallEntryPoint();
} else if (ic_data.NumArgsTested() == 2) {
- label_address = StubCode::TwoArgsUnoptimizedStaticCallEntryPoint();
+ label_address = stub_code->TwoArgsUnoptimizedStaticCallEntryPoint();
} else {
UNIMPLEMENTED();
}
@@ -1362,13 +1368,14 @@
intptr_t deopt_id,
intptr_t token_pos,
LocationSummary* locs) {
+ StubCode* stub_code = isolate()->stub_code();
__ TraceSimMsg("StaticCall");
__ LoadObject(S4, arguments_descriptor);
// Do not use the code from the function, but let the code be patched so that
// we can record the outgoing edges to other code.
GenerateDartCall(deopt_id,
token_pos,
- &StubCode::CallStaticFunctionLabel(),
+ &stub_code->CallStaticFunctionLabel(),
PcDescriptors::kOptStaticCall,
locs);
AddStaticCallTarget(function);
@@ -1382,6 +1389,7 @@
intptr_t token_pos) {
__ TraceSimMsg("EqualityRegConstCompare");
if (needs_number_check) {
+ StubCode* stub_code = isolate()->stub_code();
ASSERT(!obj.IsMint() && !obj.IsDouble() && !obj.IsBigint());
__ addiu(SP, SP, Immediate(-2 * kWordSize));
__ sw(reg, Address(SP, 1 * kWordSize));
@@ -1389,10 +1397,10 @@
__ sw(TMP, Address(SP, 0 * kWordSize));
if (is_optimizing()) {
__ BranchLinkPatchable(
- &StubCode::OptimizedIdenticalWithNumberCheckLabel());
+ &stub_code->OptimizedIdenticalWithNumberCheckLabel());
} else {
__ BranchLinkPatchable(
- &StubCode::UnoptimizedIdenticalWithNumberCheckLabel());
+ &stub_code->UnoptimizedIdenticalWithNumberCheckLabel());
}
if (token_pos != Scanner::kNoSourcePos) {
AddCurrentDescriptor(PcDescriptors::kRuntimeCall,
@@ -1415,17 +1423,18 @@
__ TraceSimMsg("EqualityRegRegCompare");
__ Comment("EqualityRegRegCompare");
if (needs_number_check) {
+ StubCode* stub_code = isolate()->stub_code();
__ addiu(SP, SP, Immediate(-2 * kWordSize));
__ sw(left, Address(SP, 1 * kWordSize));
__ sw(right, Address(SP, 0 * kWordSize));
if (is_optimizing()) {
__ BranchLinkPatchable(
- &StubCode::OptimizedIdenticalWithNumberCheckLabel());
+ &stub_code->OptimizedIdenticalWithNumberCheckLabel());
} else {
__ LoadImmediate(S4, 0);
__ LoadImmediate(S5, 0);
__ BranchLinkPatchable(
- &StubCode::UnoptimizedIdenticalWithNumberCheckLabel());
+ &stub_code->UnoptimizedIdenticalWithNumberCheckLabel());
}
if (token_pos != Scanner::kNoSourcePos) {
AddCurrentDescriptor(PcDescriptors::kRuntimeCall,
@@ -1545,6 +1554,8 @@
const Array& arguments_descriptor =
Array::ZoneHandle(ArgumentsDescriptor::New(argument_count,
argument_names));
+ StubCode* stub_code = isolate()->stub_code();
+
__ TraceSimMsg("EmitTestAndCall");
__ Comment("EmitTestAndCall");
__ LoadObject(S4, arguments_descriptor);
@@ -1560,7 +1571,7 @@
// that we can record the outgoing edges to other code.
GenerateDartCall(deopt_id,
token_index,
- &StubCode::CallStaticFunctionLabel(),
+ &stub_code->CallStaticFunctionLabel(),
PcDescriptors::kOptStaticCall,
locs);
const Function& function = *sorted[i].target;
« no previous file with comments | « runtime/vm/flow_graph_compiler_ia32.cc ('k') | runtime/vm/flow_graph_compiler_x64.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698