| OLD | NEW |
| 1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 13 matching lines...) Expand all Loading... |
| 24 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 24 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| 25 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | 25 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| 26 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 26 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| 27 | 27 |
| 28 #include "v8.h" | 28 #include "v8.h" |
| 29 | 29 |
| 30 #if defined(V8_TARGET_ARCH_IA32) | 30 #if defined(V8_TARGET_ARCH_IA32) |
| 31 | 31 |
| 32 #include "ia32/lithium-codegen-ia32.h" | 32 #include "ia32/lithium-codegen-ia32.h" |
| 33 #include "code-stubs.h" | 33 #include "code-stubs.h" |
| 34 #include "deoptimizer.h" |
| 34 #include "stub-cache.h" | 35 #include "stub-cache.h" |
| 35 | 36 |
| 36 namespace v8 { | 37 namespace v8 { |
| 37 namespace internal { | 38 namespace internal { |
| 38 | 39 |
| 39 | 40 |
| 40 // When invoking builtins, we need to record the safepoint in the middle of | 41 // When invoking builtins, we need to record the safepoint in the middle of |
| 41 // the invoke instruction sequence generated by the macro assembler. | 42 // the invoke instruction sequence generated by the macro assembler. |
| 42 class SafepointGenerator : public PostCallGenerator { | 43 class SafepointGenerator : public PostCallGenerator { |
| 43 public: | 44 public: |
| 44 SafepointGenerator(LCodeGen* codegen, | 45 SafepointGenerator(LCodeGen* codegen, |
| 45 LPointerMap* pointers, | 46 LPointerMap* pointers, |
| 46 int deoptimization_index) | 47 int deoptimization_index) |
| 47 : codegen_(codegen), | 48 : codegen_(codegen), |
| 48 pointers_(pointers), | 49 pointers_(pointers), |
| 49 deoptimization_index_(deoptimization_index) {} | 50 deoptimization_index_(deoptimization_index) {} |
| 50 virtual ~SafepointGenerator() { } | 51 virtual ~SafepointGenerator() { } |
| 51 | 52 |
| 52 virtual void Generate() { | 53 virtual void Generate() { |
| 53 // Ensure that we have enough space in the reloc info to patch | |
| 54 // this with calls when doing deoptimization. | |
| 55 codegen_->EnsureRelocSpaceForDeoptimization(); | |
| 56 codegen_->RecordSafepoint(pointers_, deoptimization_index_); | 54 codegen_->RecordSafepoint(pointers_, deoptimization_index_); |
| 57 } | 55 } |
| 58 | 56 |
| 59 private: | 57 private: |
| 60 LCodeGen* codegen_; | 58 LCodeGen* codegen_; |
| 61 LPointerMap* pointers_; | 59 LPointerMap* pointers_; |
| 62 int deoptimization_index_; | 60 int deoptimization_index_; |
| 63 }; | 61 }; |
| 64 | 62 |
| 65 | 63 |
| 66 #define __ masm()-> | 64 #define __ masm()-> |
| 67 | 65 |
| 68 bool LCodeGen::GenerateCode() { | 66 bool LCodeGen::GenerateCode() { |
| 69 HPhase phase("Code generation", chunk()); | 67 HPhase phase("Code generation", chunk()); |
| 70 ASSERT(is_unused()); | 68 ASSERT(is_unused()); |
| 71 status_ = GENERATING; | 69 status_ = GENERATING; |
| 72 CpuFeatures::Scope scope(SSE2); | 70 CpuFeatures::Scope scope(SSE2); |
| 73 return GeneratePrologue() && | 71 return GeneratePrologue() && |
| 74 GenerateBody() && | 72 GenerateBody() && |
| 75 GenerateDeferredCode() && | 73 GenerateDeferredCode() && |
| 76 GenerateRelocPadding() && | |
| 77 GenerateSafepointTable(); | 74 GenerateSafepointTable(); |
| 78 } | 75 } |
| 79 | 76 |
| 80 | 77 |
| 81 void LCodeGen::FinishCode(Handle<Code> code) { | 78 void LCodeGen::FinishCode(Handle<Code> code) { |
| 82 ASSERT(is_done()); | 79 ASSERT(is_done()); |
| 83 code->set_stack_slots(StackSlotCount()); | 80 code->set_stack_slots(StackSlotCount()); |
| 84 code->set_safepoint_table_offset(safepoints_.GetCodeOffset()); | 81 code->set_safepoint_table_offset(safepoints_.GetCodeOffset()); |
| 85 PopulateDeoptimizationData(code); | 82 PopulateDeoptimizationData(code); |
| 83 Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(code); |
| 86 } | 84 } |
| 87 | 85 |
| 88 | 86 |
| 89 void LCodeGen::Abort(const char* format, ...) { | 87 void LCodeGen::Abort(const char* format, ...) { |
| 90 if (FLAG_trace_bailout) { | 88 if (FLAG_trace_bailout) { |
| 91 SmartPointer<char> name(info()->shared_info()->DebugName()->ToCString()); | 89 SmartPointer<char> name(info()->shared_info()->DebugName()->ToCString()); |
| 92 PrintF("Aborting LCodeGen in @\"%s\": ", *name); | 90 PrintF("Aborting LCodeGen in @\"%s\": ", *name); |
| 93 va_list arguments; | 91 va_list arguments; |
| 94 va_start(arguments, format); | 92 va_start(arguments, format); |
| 95 OS::VPrint(format, arguments); | 93 OS::VPrint(format, arguments); |
| (...skipping 277 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 373 environment->spilled_double_registers()[value->index()], | 371 environment->spilled_double_registers()[value->index()], |
| 374 false); | 372 false); |
| 375 } | 373 } |
| 376 } | 374 } |
| 377 | 375 |
| 378 AddToTranslation(translation, value, environment->HasTaggedValueAt(i)); | 376 AddToTranslation(translation, value, environment->HasTaggedValueAt(i)); |
| 379 } | 377 } |
| 380 } | 378 } |
| 381 | 379 |
| 382 | 380 |
| 383 void LCodeGen::EnsureRelocSpaceForDeoptimization() { | |
| 384 // Since we patch the reloc info with RUNTIME_ENTRY calls every | |
| 385 // patch site will take up 2 bytes + any pc-jumps. We are | |
| 386 // conservative and always reserve 6 bytes in case a simple pc-jump | |
| 387 // is not enough. | |
| 388 uint32_t pc_delta = | |
| 389 masm()->pc_offset() - deoptimization_reloc_size.last_pc_offset; | |
| 390 if (is_uintn(pc_delta, 6)) { | |
| 391 deoptimization_reloc_size.min_size += 2; | |
| 392 } else { | |
| 393 deoptimization_reloc_size.min_size += 6; | |
| 394 } | |
| 395 deoptimization_reloc_size.last_pc_offset = masm()->pc_offset(); | |
| 396 } | |
| 397 | |
| 398 | |
| 399 void LCodeGen::AddToTranslation(Translation* translation, | 381 void LCodeGen::AddToTranslation(Translation* translation, |
| 400 LOperand* op, | 382 LOperand* op, |
| 401 bool is_tagged) { | 383 bool is_tagged) { |
| 402 if (op == NULL) { | 384 if (op == NULL) { |
| 403 // TODO(twuerthinger): Introduce marker operands to indicate that this value | 385 // TODO(twuerthinger): Introduce marker operands to indicate that this value |
| 404 // is not present and must be reconstructed from the deoptimizer. Currently | 386 // is not present and must be reconstructed from the deoptimizer. Currently |
| 405 // this is only used for the arguments object. | 387 // this is only used for the arguments object. |
| 406 translation->StoreArgumentsObject(); | 388 translation->StoreArgumentsObject(); |
| 407 } else if (op->IsStackSlot()) { | 389 } else if (op->IsStackSlot()) { |
| 408 if (is_tagged) { | 390 if (is_tagged) { |
| (...skipping 33 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 442 bool adjusted) { | 424 bool adjusted) { |
| 443 ASSERT(instr != NULL); | 425 ASSERT(instr != NULL); |
| 444 LPointerMap* pointers = instr->pointer_map(); | 426 LPointerMap* pointers = instr->pointer_map(); |
| 445 RecordPosition(pointers->position()); | 427 RecordPosition(pointers->position()); |
| 446 | 428 |
| 447 if (!adjusted) { | 429 if (!adjusted) { |
| 448 __ mov(esi, Operand(ebp, StandardFrameConstants::kContextOffset)); | 430 __ mov(esi, Operand(ebp, StandardFrameConstants::kContextOffset)); |
| 449 } | 431 } |
| 450 __ call(code, mode); | 432 __ call(code, mode); |
| 451 | 433 |
| 452 EnsureRelocSpaceForDeoptimization(); | |
| 453 RegisterLazyDeoptimization(instr); | 434 RegisterLazyDeoptimization(instr); |
| 454 | 435 |
| 455 // Signal that we don't inline smi code before these stubs in the | 436 // Signal that we don't inline smi code before these stubs in the |
| 456 // optimizing code generator. | 437 // optimizing code generator. |
| 457 if (code->kind() == Code::TYPE_RECORDING_BINARY_OP_IC || | 438 if (code->kind() == Code::TYPE_RECORDING_BINARY_OP_IC || |
| 458 code->kind() == Code::COMPARE_IC) { | 439 code->kind() == Code::COMPARE_IC) { |
| 459 __ nop(); | 440 __ nop(); |
| 460 } | 441 } |
| 461 } | 442 } |
| 462 | 443 |
| 463 | 444 |
| 464 void LCodeGen::CallRuntime(const Runtime::Function* fun, | 445 void LCodeGen::CallRuntime(const Runtime::Function* fun, |
| 465 int argc, | 446 int argc, |
| 466 LInstruction* instr, | 447 LInstruction* instr, |
| 467 bool adjusted) { | 448 bool adjusted) { |
| 468 ASSERT(instr != NULL); | 449 ASSERT(instr != NULL); |
| 469 ASSERT(instr->HasPointerMap()); | 450 ASSERT(instr->HasPointerMap()); |
| 470 LPointerMap* pointers = instr->pointer_map(); | 451 LPointerMap* pointers = instr->pointer_map(); |
| 471 RecordPosition(pointers->position()); | 452 RecordPosition(pointers->position()); |
| 472 | 453 |
| 473 if (!adjusted) { | 454 if (!adjusted) { |
| 474 __ mov(esi, Operand(ebp, StandardFrameConstants::kContextOffset)); | 455 __ mov(esi, Operand(ebp, StandardFrameConstants::kContextOffset)); |
| 475 } | 456 } |
| 476 __ CallRuntime(fun, argc); | 457 __ CallRuntime(fun, argc); |
| 477 EnsureRelocSpaceForDeoptimization(); | 458 |
| 478 RegisterLazyDeoptimization(instr); | 459 RegisterLazyDeoptimization(instr); |
| 479 } | 460 } |
| 480 | 461 |
| 481 | 462 |
| 482 void LCodeGen::RegisterLazyDeoptimization(LInstruction* instr) { | 463 void LCodeGen::RegisterLazyDeoptimization(LInstruction* instr) { |
| 483 // Create the environment to bailout to. If the call has side effects | 464 // Create the environment to bailout to. If the call has side effects |
| 484 // execution has to continue after the call otherwise execution can continue | 465 // execution has to continue after the call otherwise execution can continue |
| 485 // from a previous bailout point repeating the call. | 466 // from a previous bailout point repeating the call. |
| 486 LEnvironment* deoptimization_environment; | 467 LEnvironment* deoptimization_environment; |
| 487 if (instr->HasDeoptimizationEnvironment()) { | 468 if (instr->HasDeoptimizationEnvironment()) { |
| (...skipping 2006 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2494 | 2475 |
| 2495 LPointerMap* pointers = instr->pointer_map(); | 2476 LPointerMap* pointers = instr->pointer_map(); |
| 2496 RecordPosition(pointers->position()); | 2477 RecordPosition(pointers->position()); |
| 2497 | 2478 |
| 2498 // Invoke function. | 2479 // Invoke function. |
| 2499 if (*function == *info()->closure()) { | 2480 if (*function == *info()->closure()) { |
| 2500 __ CallSelf(); | 2481 __ CallSelf(); |
| 2501 } else { | 2482 } else { |
| 2502 __ call(FieldOperand(edi, JSFunction::kCodeEntryOffset)); | 2483 __ call(FieldOperand(edi, JSFunction::kCodeEntryOffset)); |
| 2503 } | 2484 } |
| 2504 EnsureRelocSpaceForDeoptimization(); | |
| 2505 | 2485 |
| 2506 // Setup deoptimization. | 2486 // Setup deoptimization. |
| 2507 RegisterLazyDeoptimization(instr); | 2487 RegisterLazyDeoptimization(instr); |
| 2508 } | 2488 } |
| 2509 | 2489 |
| 2510 | 2490 |
| 2511 void LCodeGen::DoCallConstantFunction(LCallConstantFunction* instr) { | 2491 void LCodeGen::DoCallConstantFunction(LCallConstantFunction* instr) { |
| 2512 ASSERT(ToRegister(instr->result()).is(eax)); | 2492 ASSERT(ToRegister(instr->result()).is(eax)); |
| 2513 __ mov(edi, instr->function()); | 2493 __ mov(edi, instr->function()); |
| 2514 CallKnownFunction(instr->function(), instr->arity(), instr); | 2494 CallKnownFunction(instr->function(), instr->arity(), instr); |
| (...skipping 1550 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 4065 ASSERT(osr_pc_offset_ == -1); | 4045 ASSERT(osr_pc_offset_ == -1); |
| 4066 osr_pc_offset_ = masm()->pc_offset(); | 4046 osr_pc_offset_ = masm()->pc_offset(); |
| 4067 } | 4047 } |
| 4068 | 4048 |
| 4069 | 4049 |
| 4070 #undef __ | 4050 #undef __ |
| 4071 | 4051 |
| 4072 } } // namespace v8::internal | 4052 } } // namespace v8::internal |
| 4073 | 4053 |
| 4074 #endif // V8_TARGET_ARCH_IA32 | 4054 #endif // V8_TARGET_ARCH_IA32 |
| OLD | NEW |