Chromium Code Reviews| OLD | NEW |
|---|---|
| 1 // Copyright (c) 2016, the Dart project authors. Please see the AUTHORS file | 1 // Copyright (c) 2016, the Dart project authors. Please see the AUTHORS file |
| 2 // for details. All rights reserved. Use of this source code is governed by a | 2 // for details. All rights reserved. Use of this source code is governed by a |
| 3 // BSD-style license that can be found in the LICENSE file. | 3 // BSD-style license that can be found in the LICENSE file. |
| 4 | 4 |
| 5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_DBC. | 5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_DBC. |
| 6 #if defined(TARGET_ARCH_DBC) | 6 #if defined(TARGET_ARCH_DBC) |
| 7 | 7 |
| 8 #include "vm/flow_graph_compiler.h" | 8 #include "vm/flow_graph_compiler.h" |
| 9 | 9 |
| 10 #include "vm/ast_printer.h" | 10 #include "vm/ast_printer.h" |
| (...skipping 72 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 83 | 83 |
| 84 void FlowGraphCompiler::ExitIntrinsicMode() { | 84 void FlowGraphCompiler::ExitIntrinsicMode() { |
| 85 ASSERT(intrinsic_mode()); | 85 ASSERT(intrinsic_mode()); |
| 86 intrinsic_mode_ = false; | 86 intrinsic_mode_ = false; |
| 87 } | 87 } |
| 88 | 88 |
| 89 | 89 |
| 90 RawTypedData* CompilerDeoptInfo::CreateDeoptInfo(FlowGraphCompiler* compiler, | 90 RawTypedData* CompilerDeoptInfo::CreateDeoptInfo(FlowGraphCompiler* compiler, |
| 91 DeoptInfoBuilder* builder, | 91 DeoptInfoBuilder* builder, |
| 92 const Array& deopt_table) { | 92 const Array& deopt_table) { |
| 93 UNIMPLEMENTED(); | 93 if (deopt_env_ == NULL) { |
| 94 return TypedData::null(); | 94 ++builder->current_info_number_; |
| 95 return TypedData::null(); | |
| 96 } | |
| 97 | |
| 98 intptr_t stack_height = compiler->StackSize(); | |
| 99 AllocateIncomingParametersRecursive(deopt_env_, &stack_height); | |
| 100 | |
| 101 intptr_t slot_ix = 0; | |
| 102 Environment* current = deopt_env_; | |
| 103 | |
| 104 // Emit all kMaterializeObject instructions describing objects to be | |
| 105 // materialized on the deoptimization as a prefix to the deoptimization info. | |
| 106 EmitMaterializations(deopt_env_, builder); | |
| 107 | |
| 108 // The real frame starts here. | |
| 109 builder->MarkFrameStart(); | |
| 110 | |
| 111 Zone* zone = compiler->zone(); | |
| 112 | |
| 113 builder->AddCallerFp(slot_ix++); | |
| 114 builder->AddReturnAddress(current->function(), deopt_id(), slot_ix++); | |
| 115 builder->AddPcMarker(Function::ZoneHandle(zone), slot_ix++); | |
| 116 builder->AddConstant(Function::ZoneHandle(zone), slot_ix++); | |
| 117 | |
| 118 // Emit all values that are needed for materialization as a part of the | |
| 119 // expression stack for the bottom-most frame. This guarantees that GC | |
| 120 // will be able to find them during materialization. | |
| 121 slot_ix = builder->EmitMaterializationArguments(slot_ix); | |
| 122 | |
| 123 // For the innermost environment, set outgoing arguments and the locals. | |
| 124 for (intptr_t i = current->Length() - 1; | |
| 125 i >= current->fixed_parameter_count(); | |
| 126 i--) { | |
| 127 builder->AddCopy(current->ValueAt(i), current->LocationAt(i), slot_ix++); | |
| 128 } | |
| 129 | |
|
zra
2016/05/19 16:24:27
Extra newline.
Vyacheslav Egorov (Google)
2016/05/20 12:11:47
Done.
| |
| 130 | |
| 131 builder->AddCallerFp(slot_ix++); | |
| 132 | |
| 133 Environment* previous = current; | |
| 134 current = current->outer(); | |
| 135 while (current != NULL) { | |
| 136 // For any outer environment the deopt id is that of the call instruction | |
| 137 // which is recorded in the outer environment. | |
| 138 builder->AddReturnAddress( | |
| 139 current->function(), | |
| 140 Thread::ToDeoptAfter(current->deopt_id()), | |
| 141 slot_ix++); | |
| 142 | |
| 143 builder->AddPcMarker(previous->function(), slot_ix++); | |
| 144 builder->AddConstant(previous->function(), slot_ix++); | |
| 145 | |
| 146 // The values of outgoing arguments can be changed from the inlined call so | |
| 147 // we must read them from the previous environment. | |
| 148 for (intptr_t i = previous->fixed_parameter_count() - 1; i >= 0; i--) { | |
| 149 builder->AddCopy(previous->ValueAt(i), | |
| 150 previous->LocationAt(i), | |
| 151 slot_ix++); | |
| 152 } | |
| 153 | |
| 154 // Set the locals, note that outgoing arguments are not in the environment. | |
| 155 for (intptr_t i = current->Length() - 1; | |
| 156 i >= current->fixed_parameter_count(); | |
| 157 i--) { | |
| 158 builder->AddCopy(current->ValueAt(i), | |
| 159 current->LocationAt(i), | |
| 160 slot_ix++); | |
| 161 } | |
| 162 | |
| 163 builder->AddCallerFp(slot_ix++); | |
| 164 | |
| 165 // Iterate on the outer environment. | |
| 166 previous = current; | |
| 167 current = current->outer(); | |
| 168 } | |
| 169 // The previous pointer is now the outermost environment. | |
| 170 ASSERT(previous != NULL); | |
| 171 | |
| 172 // For the outermost environment, set caller PC. | |
| 173 builder->AddCallerPc(slot_ix++); | |
| 174 | |
| 175 builder->AddPcMarker(previous->function(), slot_ix++); | |
| 176 builder->AddConstant(previous->function(), slot_ix++); | |
| 177 | |
| 178 | |
| 179 // For the outermost environment, set the incoming arguments. | |
| 180 for (intptr_t i = previous->fixed_parameter_count() - 1; i >= 0; i--) { | |
| 181 builder->AddCopy(previous->ValueAt(i), previous->LocationAt(i), slot_ix++); | |
| 182 } | |
| 183 | |
| 184 return builder->CreateDeoptInfo(deopt_table); | |
| 95 } | 185 } |
| 96 | 186 |
| 97 | 187 |
| 188 void FlowGraphCompiler::RecordAfterCall(Instruction* instr) { | |
| 189 RecordSafepoint(instr->locs()); | |
| 190 // Marks either the continuation point in unoptimized code or the | |
| 191 // deoptimization point in optimized code, after call. | |
| 192 const intptr_t deopt_id_after = Thread::ToDeoptAfter(instr->deopt_id()); | |
| 193 if (is_optimizing()) { | |
| 194 // Return/ReturnTOS instruction drops incoming arguments so | |
| 195 // we have to drop outgoing arguments from the innermost environment. | |
| 196 // On all other architectures caller drops outgoing arguments itself | |
| 197 // hence the difference. | |
| 198 pending_deoptimization_env_->DropArguments(instr->ArgumentCount()); | |
| 199 AddDeoptIndexAtCall(deopt_id_after, instr->token_pos()); | |
| 200 } else { | |
| 201 // Add deoptimization continuation point after the call and before the | |
| 202 // arguments are removed. | |
| 203 // In optimized code this descriptor is needed for exception handling. | |
| 204 AddCurrentDescriptor(RawPcDescriptors::kDeopt, | |
| 205 deopt_id_after, | |
| 206 instr->token_pos()); | |
| 207 } | |
| 208 } | |
| 209 | |
| 210 | |
| 98 void CompilerDeoptInfoWithStub::GenerateCode(FlowGraphCompiler* compiler, | 211 void CompilerDeoptInfoWithStub::GenerateCode(FlowGraphCompiler* compiler, |
| 99 intptr_t stub_ix) { | 212 intptr_t stub_ix) { |
| 100 UNIMPLEMENTED(); | 213 UNIMPLEMENTED(); |
| 101 } | 214 } |
| 102 | 215 |
| 103 | 216 |
| 104 #define __ assembler()-> | 217 #define __ assembler()-> |
| 105 | 218 |
| 106 | 219 |
| 107 void FlowGraphCompiler::GenerateAssertAssignable(TokenPosition token_pos, | 220 void FlowGraphCompiler::GenerateAssertAssignable(TokenPosition token_pos, |
| 108 intptr_t deopt_id, | 221 intptr_t deopt_id, |
| 109 const AbstractType& dst_type, | 222 const AbstractType& dst_type, |
| 110 const String& dst_name, | 223 const String& dst_name, |
| 111 LocationSummary* locs) { | 224 LocationSummary* locs) { |
| 112 ASSERT(!is_optimizing()); | |
| 113 SubtypeTestCache& test_cache = SubtypeTestCache::Handle(); | 225 SubtypeTestCache& test_cache = SubtypeTestCache::Handle(); |
| 114 if (!dst_type.IsVoidType() && dst_type.IsInstantiated()) { | 226 if (!dst_type.IsVoidType() && dst_type.IsInstantiated()) { |
| 115 test_cache = SubtypeTestCache::New(); | 227 test_cache = SubtypeTestCache::New(); |
| 116 } | 228 } |
| 117 | 229 |
| 230 if (is_optimizing()) { | |
| 231 __ Push(locs->in(0).reg()); | |
| 232 __ Push(locs->in(1).reg()); | |
| 233 } | |
| 118 __ PushConstant(dst_type); | 234 __ PushConstant(dst_type); |
| 119 __ PushConstant(dst_name); | 235 __ PushConstant(dst_name); |
| 120 __ AssertAssignable(__ AddConstant(test_cache)); | 236 __ AssertAssignable(__ AddConstant(test_cache)); |
| 237 RecordSafepoint(locs); | |
| 121 AddCurrentDescriptor(RawPcDescriptors::kOther, deopt_id, token_pos); | 238 AddCurrentDescriptor(RawPcDescriptors::kOther, deopt_id, token_pos); |
| 239 if (is_optimizing()) { | |
| 240 __ Drop1(); | |
|
Florian Schneider
2016/05/19 13:21:40
__ Drop(2)?
Vyacheslav Egorov (Google)
2016/05/19 15:19:40
Explained why Drop(1)
| |
| 241 } | |
| 122 } | 242 } |
| 123 | 243 |
| 124 | 244 |
| 125 void FlowGraphCompiler::EmitInstructionEpilogue(Instruction* instr) { | 245 void FlowGraphCompiler::EmitInstructionEpilogue(Instruction* instr) { |
| 126 if (!is_optimizing()) { | 246 if (!is_optimizing()) { |
| 127 Definition* defn = instr->AsDefinition(); | 247 Definition* defn = instr->AsDefinition(); |
| 128 if ((defn != NULL) && | 248 if ((defn != NULL) && |
| 129 (defn->tag() != Instruction::kPushArgument) && | 249 (defn->tag() != Instruction::kPushArgument) && |
| 130 (defn->tag() != Instruction::kStoreIndexed) && | 250 (defn->tag() != Instruction::kStoreIndexed) && |
| 131 (defn->tag() != Instruction::kStoreStaticField) && | 251 (defn->tag() != Instruction::kStoreStaticField) && |
| (...skipping 29 matching lines...) Expand all Loading... | |
| 161 const int num_opt_pos_params = function.NumOptionalPositionalParameters(); | 281 const int num_opt_pos_params = function.NumOptionalPositionalParameters(); |
| 162 const int num_opt_named_params = function.NumOptionalNamedParameters(); | 282 const int num_opt_named_params = function.NumOptionalNamedParameters(); |
| 163 const int num_params = | 283 const int num_params = |
| 164 num_fixed_params + num_opt_pos_params + num_opt_named_params; | 284 num_fixed_params + num_opt_pos_params + num_opt_named_params; |
| 165 const bool has_optional_params = (num_opt_pos_params != 0) || | 285 const bool has_optional_params = (num_opt_pos_params != 0) || |
| 166 (num_opt_named_params != 0); | 286 (num_opt_named_params != 0); |
| 167 const int num_locals = parsed_function().num_stack_locals(); | 287 const int num_locals = parsed_function().num_stack_locals(); |
| 168 const intptr_t context_index = | 288 const intptr_t context_index = |
| 169 -parsed_function().current_context_var()->index() - 1; | 289 -parsed_function().current_context_var()->index() - 1; |
| 170 | 290 |
| 291 if (CanOptimizeFunction() && | |
| 292 function.IsOptimizable() && | |
| 293 (!is_optimizing() || may_reoptimize())) { | |
| 294 __ HotCheck(!is_optimizing(), GetOptimizationThreshold()); | |
| 295 } | |
| 296 | |
| 171 if (has_optional_params) { | 297 if (has_optional_params) { |
| 172 __ EntryOpt(num_fixed_params, num_opt_pos_params, num_opt_named_params); | 298 __ EntryOptional(num_fixed_params, |
| 299 num_opt_pos_params, | |
| 300 num_opt_named_params); | |
| 301 } else if (!is_optimizing()) { | |
| 302 __ Entry(num_fixed_params, num_locals, context_index); | |
| 173 } else { | 303 } else { |
| 174 __ Entry(num_fixed_params, num_locals, context_index); | 304 __ EntryOptimized(num_fixed_params, |
| 305 flow_graph_.graph_entry()->spill_slot_count()); | |
| 175 } | 306 } |
| 176 | 307 |
| 177 if (num_opt_named_params != 0) { | 308 if (num_opt_named_params != 0) { |
| 178 LocalScope* scope = parsed_function().node_sequence()->scope(); | 309 LocalScope* scope = parsed_function().node_sequence()->scope(); |
| 179 | 310 |
| 180 // Start by alphabetically sorting the names of the optional parameters. | 311 // Start by alphabetically sorting the names of the optional parameters. |
| 181 LocalVariable** opt_param = | 312 LocalVariable** opt_param = |
| 182 zone()->Alloc<LocalVariable*>(num_opt_named_params); | 313 zone()->Alloc<LocalVariable*>(num_opt_named_params); |
| 183 int* opt_param_position = zone()->Alloc<int>(num_opt_named_params); | 314 int* opt_param_position = zone()->Alloc<int>(num_opt_named_params); |
| 184 for (int pos = num_fixed_params; pos < num_params; pos++) { | 315 for (int pos = num_fixed_params; pos < num_params; pos++) { |
| (...skipping 20 matching lines...) Expand all Loading... | |
| 205 __ LoadConstant(param_pos, value); | 336 __ LoadConstant(param_pos, value); |
| 206 } | 337 } |
| 207 } else if (num_opt_pos_params != 0) { | 338 } else if (num_opt_pos_params != 0) { |
| 208 for (intptr_t i = 0; i < num_opt_pos_params; i++) { | 339 for (intptr_t i = 0; i < num_opt_pos_params; i++) { |
| 209 const Object& value = parsed_function().DefaultParameterValueAt(i); | 340 const Object& value = parsed_function().DefaultParameterValueAt(i); |
| 210 __ LoadConstant(num_fixed_params + i, value); | 341 __ LoadConstant(num_fixed_params + i, value); |
| 211 } | 342 } |
| 212 } | 343 } |
| 213 | 344 |
| 214 | 345 |
| 215 ASSERT(num_locals > 0); // There is always at least context_var. | |
| 216 if (has_optional_params) { | 346 if (has_optional_params) { |
| 217 ASSERT(!is_optimizing()); | 347 if (!is_optimizing()) { |
| 218 __ Frame(num_locals); // Reserve space for locals. | 348 ASSERT(num_locals > 0); // There is always at least context_var. |
| 349 __ Frame(num_locals); // Reserve space for locals. | |
| 350 } else if (flow_graph_.graph_entry()->spill_slot_count() > | |
| 351 flow_graph_.num_copied_params()) { | |
| 352 __ Frame(flow_graph_.graph_entry()->spill_slot_count() - | |
| 353 flow_graph_.num_copied_params()); | |
| 354 } | |
| 219 } | 355 } |
| 220 | 356 |
| 221 if (function.IsClosureFunction()) { | 357 if (function.IsClosureFunction()) { |
| 222 Register reg = context_index; | 358 Register reg = is_optimizing() ? flow_graph_.num_copied_params() |
| 359 : context_index; | |
| 223 Register closure_reg = reg; | 360 Register closure_reg = reg; |
| 224 LocalScope* scope = parsed_function().node_sequence()->scope(); | 361 LocalScope* scope = parsed_function().node_sequence()->scope(); |
| 225 LocalVariable* local = scope->VariableAt(0); | 362 LocalVariable* local = scope->VariableAt(0); |
| 226 if (local->index() > 0) { | 363 if (local->index() > 0) { |
| 227 __ Move(reg, -local->index()); | 364 __ Move(reg, -local->index()); |
| 228 } else { | 365 } else { |
| 229 closure_reg = -local->index() - 1; | 366 closure_reg = -local->index() - 1; |
| 230 } | 367 } |
| 231 __ LoadField(reg, closure_reg, Closure::context_offset() / kWordSize); | 368 __ LoadField(reg, closure_reg, Closure::context_offset() / kWordSize); |
| 232 } else if (has_optional_params) { | 369 } else if (has_optional_params && !is_optimizing()) { |
| 233 __ LoadConstant(context_index, | 370 __ LoadConstant(context_index, |
| 234 Object::Handle(isolate()->object_store()->empty_context())); | 371 Object::Handle(isolate()->object_store()->empty_context())); |
| 235 } | 372 } |
| 236 } | 373 } |
| 237 | 374 |
| 238 | 375 |
| 239 void FlowGraphCompiler::CompileGraph() { | 376 void FlowGraphCompiler::CompileGraph() { |
| 240 InitCompiler(); | 377 InitCompiler(); |
| 241 | 378 |
| 242 if (TryIntrinsify()) { | 379 if (TryIntrinsify()) { |
| 243 // Skip regular code generation. | 380 // Skip regular code generation. |
| 244 return; | 381 return; |
| 245 } | 382 } |
| 246 | 383 |
| 247 EmitFrameEntry(); | 384 EmitFrameEntry(); |
| 248 VisitBlocks(); | 385 VisitBlocks(); |
| 249 } | 386 } |
| 250 | 387 |
| 251 | 388 |
| 252 #undef __ | 389 #undef __ |
| 253 #define __ compiler_->assembler()-> | 390 #define __ compiler_->assembler()-> |
| 254 | 391 |
| 255 | 392 |
| 256 void ParallelMoveResolver::EmitMove(int index) { | 393 void ParallelMoveResolver::EmitMove(int index) { |
| 257 UNIMPLEMENTED(); | 394 MoveOperands* move = moves_[index]; |
| 395 const Location source = move->src(); | |
| 396 const Location destination = move->dest(); | |
| 397 if (source.IsStackSlot() && destination.IsRegister()) { | |
| 398 // Only allow access to the arguments. | |
| 399 ASSERT(source.base_reg() == FPREG); | |
| 400 ASSERT(source.stack_index() < 0); | |
| 401 __ Move(destination.reg(), -kParamEndSlotFromFp + source.stack_index()); | |
| 402 } else if (source.IsRegister() && destination.IsRegister()) { | |
| 403 __ Move(destination.reg(), source.reg()); | |
| 404 } else if (source.IsConstant() && destination.IsRegister()) { | |
| 405 __ LoadConstant(destination.reg(), source.constant()); | |
| 406 } else { | |
| 407 compiler_->Bailout("Unsupported move"); | |
| 408 } | |
| 409 | |
| 410 move->Eliminate(); | |
| 258 } | 411 } |
| 259 | 412 |
| 260 | 413 |
| 261 void ParallelMoveResolver::EmitSwap(int index) { | 414 void ParallelMoveResolver::EmitSwap(int index) { |
| 262 UNIMPLEMENTED(); | 415 MoveOperands* move = moves_[index]; |
| 416 const Location source = move->src(); | |
| 417 const Location destination = move->dest(); | |
| 418 | |
| 419 if (source.IsRegister() && destination.IsRegister()) { | |
| 420 __ Swap(destination.reg(), source.reg()); | |
| 421 } else { | |
| 422 UNREACHABLE(); | |
|
Florian Schneider
2016/05/19 13:21:40
Use ASSERT instead of if-statement?
Vyacheslav Egorov (Google)
2016/05/19 15:19:40
Done.
| |
| 423 } | |
| 424 | |
| 425 // The swap of source and destination has executed a move from source to | |
| 426 // destination. | |
| 427 move->Eliminate(); | |
| 428 | |
| 429 // Any unperformed (including pending) move with a source of either | |
| 430 // this move's source or destination needs to have their source | |
| 431 // changed to reflect the state of affairs after the swap. | |
| 432 for (int i = 0; i < moves_.length(); ++i) { | |
| 433 const MoveOperands& other_move = *moves_[i]; | |
| 434 if (other_move.Blocks(source)) { | |
| 435 moves_[i]->set_src(destination); | |
| 436 } else if (other_move.Blocks(destination)) { | |
| 437 moves_[i]->set_src(source); | |
| 438 } | |
| 439 } | |
| 263 } | 440 } |
| 264 | 441 |
| 265 | 442 |
| 266 void ParallelMoveResolver::MoveMemoryToMemory(const Address& dst, | 443 void ParallelMoveResolver::MoveMemoryToMemory(const Address& dst, |
| 267 const Address& src) { | 444 const Address& src) { |
| 268 UNREACHABLE(); | 445 UNREACHABLE(); |
| 269 } | 446 } |
| 270 | 447 |
| 271 | 448 |
| 272 void ParallelMoveResolver::StoreObject(const Address& dst, const Object& obj) { | 449 void ParallelMoveResolver::StoreObject(const Address& dst, const Object& obj) { |
| (...skipping 48 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 321 void ParallelMoveResolver::RestoreFpuScratch(FpuRegister reg) { | 498 void ParallelMoveResolver::RestoreFpuScratch(FpuRegister reg) { |
| 322 UNIMPLEMENTED(); | 499 UNIMPLEMENTED(); |
| 323 } | 500 } |
| 324 | 501 |
| 325 | 502 |
| 326 #undef __ | 503 #undef __ |
| 327 | 504 |
| 328 } // namespace dart | 505 } // namespace dart |
| 329 | 506 |
| 330 #endif // defined TARGET_ARCH_DBC | 507 #endif // defined TARGET_ARCH_DBC |
| OLD | NEW |