Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(159)

Side by Side Diff: src/ia32/lithium-ia32.cc

Issue 10701054: Enable stub generation using Hydrogen/Lithium (again) (Closed) Base URL: https://v8.googlecode.com/svn/branches/bleeding_edge
Patch Set: First pass at pre-VFP2 RA Created 8 years, 1 month ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch | Annotate | Revision Log
OLDNEW
1 // Copyright 2012 the V8 project authors. All rights reserved. 1 // Copyright 2012 the V8 project authors. All rights reserved.
2 // Redistribution and use in source and binary forms, with or without 2 // Redistribution and use in source and binary forms, with or without
3 // modification, are permitted provided that the following conditions are 3 // modification, are permitted provided that the following conditions are
4 // met: 4 // met:
5 // 5 //
6 // * Redistributions of source code must retain the above copyright 6 // * Redistributions of source code must retain the above copyright
7 // notice, this list of conditions and the following disclaimer. 7 // notice, this list of conditions and the following disclaimer.
8 // * Redistributions in binary form must reproduce the above 8 // * Redistributions in binary form must reproduce the above
9 // copyright notice, this list of conditions and the following 9 // copyright notice, this list of conditions and the following
10 // disclaimer in the documentation and/or other materials provided 10 // disclaimer in the documentation and/or other materials provided
(...skipping 26 matching lines...) Expand all
37 namespace internal { 37 namespace internal {
38 38
39 #define DEFINE_COMPILE(type) \ 39 #define DEFINE_COMPILE(type) \
40 void L##type::CompileToNative(LCodeGen* generator) { \ 40 void L##type::CompileToNative(LCodeGen* generator) { \
41 generator->Do##type(this); \ 41 generator->Do##type(this); \
42 } 42 }
43 LITHIUM_CONCRETE_INSTRUCTION_LIST(DEFINE_COMPILE) 43 LITHIUM_CONCRETE_INSTRUCTION_LIST(DEFINE_COMPILE)
44 #undef DEFINE_COMPILE 44 #undef DEFINE_COMPILE
45 45
46 LOsrEntry::LOsrEntry() { 46 LOsrEntry::LOsrEntry() {
47 for (int i = 0; i < Register::kNumAllocatableRegisters; ++i) { 47 for (int i = 0; i < Register::NumAllocatableRegisters(); ++i) {
48 register_spills_[i] = NULL; 48 register_spills_[i] = NULL;
49 } 49 }
50 for (int i = 0; i < DoubleRegister::kNumAllocatableRegisters; ++i) { 50 for (int i = 0; i < DoubleRegister::NumAllocatableRegisters(); ++i) {
51 double_register_spills_[i] = NULL; 51 double_register_spills_[i] = NULL;
52 } 52 }
53 } 53 }
54 54
55 55
56 void LOsrEntry::MarkSpilledRegister(int allocation_index, 56 void LOsrEntry::MarkSpilledRegister(int allocation_index,
57 LOperand* spill_operand) { 57 LOperand* spill_operand) {
58 ASSERT(spill_operand->IsStackSlot()); 58 ASSERT(spill_operand->IsStackSlot());
59 ASSERT(register_spills_[allocation_index] == NULL); 59 ASSERT(register_spills_[allocation_index] == NULL);
60 register_spills_[allocation_index] = spill_operand; 60 register_spills_[allocation_index] = spill_operand;
(...skipping 371 matching lines...) Expand 10 before | Expand all | Expand 10 after
432 } 432 }
433 433
434 434
435 LPlatformChunk* LChunkBuilder::Build() { 435 LPlatformChunk* LChunkBuilder::Build() {
436 ASSERT(is_unused()); 436 ASSERT(is_unused());
437 chunk_ = new(zone()) LPlatformChunk(info(), graph()); 437 chunk_ = new(zone()) LPlatformChunk(info(), graph());
438 HPhase phase("L_Building chunk", chunk_); 438 HPhase phase("L_Building chunk", chunk_);
439 status_ = BUILDING; 439 status_ = BUILDING;
440 440
441 // Reserve the first spill slot for the state of dynamic alignment. 441 // Reserve the first spill slot for the state of dynamic alignment.
442 int alignment_state_index = chunk_->GetNextSpillIndex(false); 442 if (info()->IsOptimizing()) {
443 ASSERT_EQ(alignment_state_index, 0); 443 int alignment_state_index = chunk_->GetNextSpillIndex(false);
444 USE(alignment_state_index); 444 ASSERT_EQ(alignment_state_index, 0);
445 USE(alignment_state_index);
446 }
445 447
446 const ZoneList<HBasicBlock*>* blocks = graph()->blocks(); 448 const ZoneList<HBasicBlock*>* blocks = graph()->blocks();
447 for (int i = 0; i < blocks->length(); i++) { 449 for (int i = 0; i < blocks->length(); i++) {
448 HBasicBlock* next = NULL; 450 HBasicBlock* next = NULL;
449 if (i < blocks->length() - 1) next = blocks->at(i + 1); 451 if (i < blocks->length() - 1) next = blocks->at(i + 1);
450 DoBasicBlock(blocks->at(i), next); 452 DoBasicBlock(blocks->at(i), next);
451 if (is_aborted()) return NULL; 453 if (is_aborted()) return NULL;
452 } 454 }
453 status_ = DONE; 455 status_ = DONE;
454 return chunk_; 456 return chunk_;
(...skipping 11 matching lines...) Expand all
466 Register::ToAllocationIndex(reg)); 468 Register::ToAllocationIndex(reg));
467 } 469 }
468 470
469 471
470 LUnallocated* LChunkBuilder::ToUnallocated(XMMRegister reg) { 472 LUnallocated* LChunkBuilder::ToUnallocated(XMMRegister reg) {
471 return new(zone()) LUnallocated(LUnallocated::FIXED_DOUBLE_REGISTER, 473 return new(zone()) LUnallocated(LUnallocated::FIXED_DOUBLE_REGISTER,
472 XMMRegister::ToAllocationIndex(reg)); 474 XMMRegister::ToAllocationIndex(reg));
473 } 475 }
474 476
475 477
478 LUnallocated* LChunkBuilder::ToUnallocated(X87TopOfStackProxyRegister reg) {
479 return new(zone()) LUnallocated(LUnallocated::FIXED_DOUBLE_REGISTER,
480 X87TopOfStackProxyRegister::ToAllocationIndex(reg));
481 }
482
483
476 LOperand* LChunkBuilder::UseFixed(HValue* value, Register fixed_register) { 484 LOperand* LChunkBuilder::UseFixed(HValue* value, Register fixed_register) {
477 return Use(value, ToUnallocated(fixed_register)); 485 return Use(value, ToUnallocated(fixed_register));
478 } 486 }
479 487
480 488
481 LOperand* LChunkBuilder::UseFixedDouble(HValue* value, XMMRegister reg) { 489 LOperand* LChunkBuilder::UseFixedDouble(HValue* value, XMMRegister reg) {
482 return Use(value, ToUnallocated(reg)); 490 return Use(value, ToUnallocated(reg));
483 } 491 }
484 492
485 493
(...skipping 112 matching lines...) Expand 10 before | Expand all | Expand 10 after
598 606
599 607
600 template<int I, int T> 608 template<int I, int T>
601 LInstruction* LChunkBuilder::DefineFixedDouble( 609 LInstruction* LChunkBuilder::DefineFixedDouble(
602 LTemplateInstruction<1, I, T>* instr, 610 LTemplateInstruction<1, I, T>* instr,
603 XMMRegister reg) { 611 XMMRegister reg) {
604 return Define(instr, ToUnallocated(reg)); 612 return Define(instr, ToUnallocated(reg));
605 } 613 }
606 614
607 615
616 template<int I, int T>
617 LInstruction* LChunkBuilder::DefineX87TOS(
618 LTemplateInstruction<1, I, T>* instr) {
619 return Define(instr, ToUnallocated(x87tos));
620 }
621
622
608 LInstruction* LChunkBuilder::AssignEnvironment(LInstruction* instr) { 623 LInstruction* LChunkBuilder::AssignEnvironment(LInstruction* instr) {
609 HEnvironment* hydrogen_env = current_block_->last_environment(); 624 HEnvironment* hydrogen_env = current_block_->last_environment();
610 int argument_index_accumulator = 0; 625 int argument_index_accumulator = 0;
611 instr->set_environment(CreateEnvironment(hydrogen_env, 626 instr->set_environment(CreateEnvironment(hydrogen_env,
612 &argument_index_accumulator)); 627 &argument_index_accumulator));
613 return instr; 628 return instr;
614 } 629 }
615 630
616 631
617 LInstruction* LChunkBuilder::MarkAsCall(LInstruction* instr, 632 LInstruction* LChunkBuilder::MarkAsCall(LInstruction* instr,
618 HInstruction* hinstr, 633 HInstruction* hinstr,
619 CanDeoptimize can_deoptimize) { 634 CanDeoptimize can_deoptimize) {
635 info()->MarkAsNonDeferredCalling();
636
620 #ifdef DEBUG 637 #ifdef DEBUG
621 instr->VerifyCall(); 638 instr->VerifyCall();
622 #endif 639 #endif
623 instr->MarkAsCall(); 640 instr->MarkAsCall();
624 instr = AssignPointerMap(instr); 641 instr = AssignPointerMap(instr);
625 642
626 if (hinstr->HasObservableSideEffects()) { 643 if (hinstr->HasObservableSideEffects()) {
627 ASSERT(hinstr->next()->IsSimulate()); 644 ASSERT(hinstr->next()->IsSimulate());
628 HSimulate* sim = HSimulate::cast(hinstr->next()); 645 HSimulate* sim = HSimulate::cast(hinstr->next());
629 ASSERT(instruction_pending_deoptimization_environment_ == NULL); 646 ASSERT(instruction_pending_deoptimization_environment_ == NULL);
(...skipping 145 matching lines...) Expand 10 before | Expand all | Expand 10 after
775 LArithmeticT* result = 792 LArithmeticT* result =
776 new(zone()) LArithmeticT(op, context, left_operand, right_operand); 793 new(zone()) LArithmeticT(op, context, left_operand, right_operand);
777 return MarkAsCall(DefineFixed(result, eax), instr); 794 return MarkAsCall(DefineFixed(result, eax), instr);
778 } 795 }
779 796
780 797
781 void LChunkBuilder::DoBasicBlock(HBasicBlock* block, HBasicBlock* next_block) { 798 void LChunkBuilder::DoBasicBlock(HBasicBlock* block, HBasicBlock* next_block) {
782 ASSERT(is_building()); 799 ASSERT(is_building());
783 current_block_ = block; 800 current_block_ = block;
784 next_block_ = next_block; 801 next_block_ = next_block;
785 if (block->IsStartBlock()) { 802 if (graph()->info()->IsOptimizing()) {
786 block->UpdateEnvironment(graph_->start_environment()); 803 if (block->IsStartBlock()) {
787 argument_count_ = 0; 804 block->UpdateEnvironment(graph_->start_environment());
788 } else if (block->predecessors()->length() == 1) { 805 argument_count_ = 0;
789 // We have a single predecessor => copy environment and outgoing 806 } else if (block->predecessors()->length() == 1) {
790 // argument count from the predecessor. 807 // We have a single predecessor => copy environment and outgoing
791 ASSERT(block->phis()->length() == 0); 808 // argument count from the predecessor.
792 HBasicBlock* pred = block->predecessors()->at(0); 809 ASSERT(block->phis()->length() == 0);
793 HEnvironment* last_environment = pred->last_environment(); 810 HBasicBlock* pred = block->predecessors()->at(0);
794 ASSERT(last_environment != NULL); 811 HEnvironment* last_environment = pred->last_environment();
795 // Only copy the environment, if it is later used again. 812 ASSERT(last_environment != NULL);
796 if (pred->end()->SecondSuccessor() == NULL) { 813 // Only copy the environment, if it is later used again.
797 ASSERT(pred->end()->FirstSuccessor() == block); 814 if (pred->end()->SecondSuccessor() == NULL) {
815 ASSERT(pred->end()->FirstSuccessor() == block);
816 } else {
817 if (pred->end()->FirstSuccessor()->block_id() > block->block_id() ||
818 pred->end()->SecondSuccessor()->block_id() > block->block_id()) {
819 last_environment = last_environment->Copy();
820 }
821 }
822 block->UpdateEnvironment(last_environment);
823 ASSERT(pred->argument_count() >= 0);
824 argument_count_ = pred->argument_count();
798 } else { 825 } else {
799 if (pred->end()->FirstSuccessor()->block_id() > block->block_id() || 826 // We are at a state join => process phis.
800 pred->end()->SecondSuccessor()->block_id() > block->block_id()) { 827 HBasicBlock* pred = block->predecessors()->at(0);
801 last_environment = last_environment->Copy(); 828 // No need to copy the environment, it cannot be used later.
829 HEnvironment* last_environment = pred->last_environment();
830 for (int i = 0; i < block->phis()->length(); ++i) {
831 HPhi* phi = block->phis()->at(i);
832 last_environment->SetValueAt(phi->merged_index(), phi);
802 } 833 }
834 for (int i = 0; i < block->deleted_phis()->length(); ++i) {
835 last_environment->SetValueAt(block->deleted_phis()->at(i),
836 graph_->GetConstantUndefined());
837 }
838 block->UpdateEnvironment(last_environment);
839 // Pick up the outgoing argument count of one of the predecessors.
840 argument_count_ = pred->argument_count();
803 } 841 }
804 block->UpdateEnvironment(last_environment);
805 ASSERT(pred->argument_count() >= 0);
806 argument_count_ = pred->argument_count();
807 } else {
808 // We are at a state join => process phis.
809 HBasicBlock* pred = block->predecessors()->at(0);
810 // No need to copy the environment, it cannot be used later.
811 HEnvironment* last_environment = pred->last_environment();
812 for (int i = 0; i < block->phis()->length(); ++i) {
813 HPhi* phi = block->phis()->at(i);
814 last_environment->SetValueAt(phi->merged_index(), phi);
815 }
816 for (int i = 0; i < block->deleted_phis()->length(); ++i) {
817 last_environment->SetValueAt(block->deleted_phis()->at(i),
818 graph_->GetConstantUndefined());
819 }
820 block->UpdateEnvironment(last_environment);
821 // Pick up the outgoing argument count of one of the predecessors.
822 argument_count_ = pred->argument_count();
823 } 842 }
824 HInstruction* current = block->first(); 843 HInstruction* current = block->first();
825 int start = chunk_->instructions()->length(); 844 int start = chunk_->instructions()->length();
826 while (current != NULL && !is_aborted()) { 845 while (current != NULL && !is_aborted()) {
827 // Code for constants in registers is generated lazily. 846 // Code for constants in registers is generated lazily.
828 if (!current->EmitAtUses()) { 847 if (!current->EmitAtUses()) {
829 VisitInstruction(current); 848 VisitInstruction(current);
830 } 849 }
831 current = current->next(); 850 current = current->next();
832 } 851 }
(...skipping 811 matching lines...) Expand 10 before | Expand all | Expand 10 after
1644 // All HForceRepresentation instructions should be eliminated in the 1663 // All HForceRepresentation instructions should be eliminated in the
1645 // representation change phase of Hydrogen. 1664 // representation change phase of Hydrogen.
1646 UNREACHABLE(); 1665 UNREACHABLE();
1647 return NULL; 1666 return NULL;
1648 } 1667 }
1649 1668
1650 1669
1651 LInstruction* LChunkBuilder::DoChange(HChange* instr) { 1670 LInstruction* LChunkBuilder::DoChange(HChange* instr) {
1652 Representation from = instr->from(); 1671 Representation from = instr->from();
1653 Representation to = instr->to(); 1672 Representation to = instr->to();
1673 // Only mark conversions that might need to allocate as calling rather than
1674 // all changes. This makes simple, non-allocating conversion not have to force
1675 // building a stack frame.
1654 if (from.IsTagged()) { 1676 if (from.IsTagged()) {
1655 if (to.IsDouble()) { 1677 if (to.IsDouble()) {
1678 info()->MarkAsDeferredCalling();
1656 LOperand* value = UseRegister(instr->value()); 1679 LOperand* value = UseRegister(instr->value());
1657 // Temp register only necessary for minus zero check. 1680 // Temp register only necessary for minus zero check.
1658 LOperand* temp = instr->deoptimize_on_minus_zero() 1681 LOperand* temp = instr->deoptimize_on_minus_zero()
1659 ? TempRegister() 1682 ? TempRegister()
1660 : NULL; 1683 : NULL;
1661 LNumberUntagD* res = new(zone()) LNumberUntagD(value, temp); 1684 LNumberUntagD* res = new(zone()) LNumberUntagD(value, temp);
1662 return AssignEnvironment(DefineAsRegister(res)); 1685 return AssignEnvironment(DefineAsRegister(res));
1663 } else { 1686 } else {
1664 ASSERT(to.IsInteger32()); 1687 ASSERT(to.IsInteger32());
1665 LOperand* value = UseRegister(instr->value()); 1688 LOperand* value = UseRegister(instr->value());
1666 if (instr->value()->type().IsSmi()) { 1689 if (instr->value()->type().IsSmi()) {
1667 return DefineSameAsFirst(new(zone()) LSmiUntag(value, false)); 1690 return DefineSameAsFirst(new(zone()) LSmiUntag(value, false));
1668 } else { 1691 } else {
1669 bool truncating = instr->CanTruncateToInt32(); 1692 bool truncating = instr->CanTruncateToInt32();
1670 LOperand* xmm_temp = 1693 LOperand* xmm_temp =
1671 (truncating && CpuFeatures::IsSupported(SSE3)) 1694 (truncating && CpuFeatures::IsSupported(SSE3))
1672 ? NULL 1695 ? NULL
1673 : FixedTemp(xmm1); 1696 : FixedTemp(xmm1);
1674 LTaggedToI* res = new(zone()) LTaggedToI(value, xmm_temp); 1697 LTaggedToI* res = new(zone()) LTaggedToI(value, xmm_temp);
1675 return AssignEnvironment(DefineSameAsFirst(res)); 1698 return AssignEnvironment(DefineSameAsFirst(res));
1676 } 1699 }
1677 } 1700 }
1678 } else if (from.IsDouble()) { 1701 } else if (from.IsDouble()) {
1679 if (to.IsTagged()) { 1702 if (to.IsTagged()) {
1680 LOperand* value = UseRegister(instr->value()); 1703 info()->MarkAsDeferredCalling();
1704 LOperand* value = CpuFeatures::IsSupported(SSE2)
1705 ? UseRegisterAtStart(instr->value())
1706 : UseAtStart(instr->value());
1681 LOperand* temp = TempRegister(); 1707 LOperand* temp = TempRegister();
1682 1708
1683 // Make sure that temp and result_temp are different registers. 1709 // Make sure that temp and result_temp are different registers.
1684 LUnallocated* result_temp = TempRegister(); 1710 LUnallocated* result_temp = TempRegister();
1685 LNumberTagD* result = new(zone()) LNumberTagD(value, temp); 1711 LNumberTagD* result = new(zone()) LNumberTagD(value, temp);
1686 return AssignPointerMap(Define(result, result_temp)); 1712 return AssignPointerMap(Define(result, result_temp));
1687 } else { 1713 } else {
1688 ASSERT(to.IsInteger32()); 1714 ASSERT(to.IsInteger32());
1689 bool truncating = instr->CanTruncateToInt32(); 1715 bool truncating = instr->CanTruncateToInt32();
1690 bool needs_temp = truncating && !CpuFeatures::IsSupported(SSE3); 1716 bool needs_temp = truncating && !CpuFeatures::IsSupported(SSE3);
1691 LOperand* value = needs_temp ? 1717 LOperand* value = needs_temp ?
1692 UseTempRegister(instr->value()) : UseRegister(instr->value()); 1718 UseTempRegister(instr->value()) : UseRegister(instr->value());
1693 LOperand* temp = needs_temp ? TempRegister() : NULL; 1719 LOperand* temp = needs_temp ? TempRegister() : NULL;
1694 return AssignEnvironment( 1720 return AssignEnvironment(
1695 DefineAsRegister(new(zone()) LDoubleToI(value, temp))); 1721 DefineAsRegister(new(zone()) LDoubleToI(value, temp)));
1696 } 1722 }
1697 } else if (from.IsInteger32()) { 1723 } else if (from.IsInteger32()) {
1724 info()->MarkAsDeferredCalling();
1698 if (to.IsTagged()) { 1725 if (to.IsTagged()) {
1699 HValue* val = instr->value(); 1726 HValue* val = instr->value();
1700 LOperand* value = UseRegister(val); 1727 LOperand* value = UseRegister(val);
1701 if (val->HasRange() && val->range()->IsInSmiRange()) { 1728 if (val->HasRange() && val->range()->IsInSmiRange()) {
1702 return DefineSameAsFirst(new(zone()) LSmiTag(value)); 1729 return DefineSameAsFirst(new(zone()) LSmiTag(value));
1703 } else if (val->CheckFlag(HInstruction::kUint32)) { 1730 } else if (val->CheckFlag(HInstruction::kUint32)) {
1704 LOperand* temp = FixedTemp(xmm1); 1731 LOperand* temp = FixedTemp(xmm1);
1705 LNumberTagU* result = new(zone()) LNumberTagU(value, temp); 1732 LNumberTagU* result = new(zone()) LNumberTagU(value, temp);
1706 return AssignEnvironment(AssignPointerMap(DefineSameAsFirst(result))); 1733 return AssignEnvironment(AssignPointerMap(DefineSameAsFirst(result)));
1707 } else { 1734 } else {
(...skipping 501 matching lines...) Expand 10 before | Expand all | Expand 10 after
2209 2236
2210 LInstruction* LChunkBuilder::DoOsrEntry(HOsrEntry* instr) { 2237 LInstruction* LChunkBuilder::DoOsrEntry(HOsrEntry* instr) {
2211 ASSERT(argument_count_ == 0); 2238 ASSERT(argument_count_ == 0);
2212 allocator_->MarkAsOsrEntry(); 2239 allocator_->MarkAsOsrEntry();
2213 current_block_->last_environment()->set_ast_id(instr->ast_id()); 2240 current_block_->last_environment()->set_ast_id(instr->ast_id());
2214 return AssignEnvironment(new(zone()) LOsrEntry); 2241 return AssignEnvironment(new(zone()) LOsrEntry);
2215 } 2242 }
2216 2243
2217 2244
2218 LInstruction* LChunkBuilder::DoParameter(HParameter* instr) { 2245 LInstruction* LChunkBuilder::DoParameter(HParameter* instr) {
2219 int spill_index = chunk()->GetParameterStackSlot(instr->index()); 2246 LParameter* result = new(zone()) LParameter;
2220 return DefineAsSpilled(new(zone()) LParameter, spill_index); 2247 switch (instr->kind()) {
2248 case FUNCTION_PARAMETER: {
2249 int spill_index = chunk()->GetParameterStackSlot(instr->index());
2250 return DefineAsSpilled(result, spill_index);
2251 }
2252 case KEYED_LOAD_IC_PARAMETER: {
2253 if (instr->index() == 0) return DefineFixed(result, edx);
2254 if (instr->index() == 1) return DefineFixed(result, ecx);
2255 UNREACHABLE();
2256 break;
2257 }
2258 case KEYED_STORE_IC_PARAMETER:
2259 if (instr->index() == 0) return DefineFixed(result, edx);
2260 if (instr->index() == 1) return DefineFixed(result, ecx);
2261 if (instr->index() == 2) return DefineFixed(result, eax);
2262 break;
2263 default:
2264 UNREACHABLE();
2265 }
2266 return NULL;
2221 } 2267 }
2222 2268
2223 2269
2224 LInstruction* LChunkBuilder::DoUnknownOSRValue(HUnknownOSRValue* instr) { 2270 LInstruction* LChunkBuilder::DoUnknownOSRValue(HUnknownOSRValue* instr) {
2225 int spill_index = chunk()->GetNextSpillIndex(false); // Not double-width. 2271 int spill_index = chunk()->GetNextSpillIndex(false); // Not double-width.
2226 if (spill_index > LUnallocated::kMaxFixedIndex) { 2272 if (spill_index > LUnallocated::kMaxFixedIndex) {
2227 Abort("Too many spill slots needed for OSR"); 2273 Abort("Too many spill slots needed for OSR");
2228 spill_index = 0; 2274 spill_index = 0;
2229 } 2275 }
2230 return DefineAsSpilled(new(zone()) LUnknownOSRValue, spill_index); 2276 return DefineAsSpilled(new(zone()) LUnknownOSRValue, spill_index);
(...skipping 80 matching lines...) Expand 10 before | Expand all | Expand 10 after
2311 instruction_pending_deoptimization_environment_ = NULL; 2357 instruction_pending_deoptimization_environment_ = NULL;
2312 pending_deoptimization_ast_id_ = BailoutId::None(); 2358 pending_deoptimization_ast_id_ = BailoutId::None();
2313 return result; 2359 return result;
2314 } 2360 }
2315 2361
2316 return NULL; 2362 return NULL;
2317 } 2363 }
2318 2364
2319 2365
2320 LInstruction* LChunkBuilder::DoStackCheck(HStackCheck* instr) { 2366 LInstruction* LChunkBuilder::DoStackCheck(HStackCheck* instr) {
2367 info()->MarkAsDeferredCalling();
2321 if (instr->is_function_entry()) { 2368 if (instr->is_function_entry()) {
2322 LOperand* context = UseFixed(instr->context(), esi); 2369 LOperand* context = UseFixed(instr->context(), esi);
2323 return MarkAsCall(new(zone()) LStackCheck(context), instr); 2370 return MarkAsCall(new(zone()) LStackCheck(context), instr);
2324 } else { 2371 } else {
2325 ASSERT(instr->is_backwards_branch()); 2372 ASSERT(instr->is_backwards_branch());
2326 LOperand* context = UseAny(instr->context()); 2373 LOperand* context = UseAny(instr->context());
2327 return AssignEnvironment( 2374 return AssignEnvironment(
2328 AssignPointerMap(new(zone()) LStackCheck(context))); 2375 AssignPointerMap(new(zone()) LStackCheck(context)));
2329 } 2376 }
2330 } 2377 }
(...skipping 70 matching lines...) Expand 10 before | Expand all | Expand 10 after
2401 LInstruction* LChunkBuilder::DoLoadFieldByIndex(HLoadFieldByIndex* instr) { 2448 LInstruction* LChunkBuilder::DoLoadFieldByIndex(HLoadFieldByIndex* instr) {
2402 LOperand* object = UseRegister(instr->object()); 2449 LOperand* object = UseRegister(instr->object());
2403 LOperand* index = UseTempRegister(instr->index()); 2450 LOperand* index = UseTempRegister(instr->index());
2404 return DefineSameAsFirst(new(zone()) LLoadFieldByIndex(object, index)); 2451 return DefineSameAsFirst(new(zone()) LLoadFieldByIndex(object, index));
2405 } 2452 }
2406 2453
2407 2454
2408 } } // namespace v8::internal 2455 } } // namespace v8::internal
2409 2456
2410 #endif // V8_TARGET_ARCH_IA32 2457 #endif // V8_TARGET_ARCH_IA32
OLDNEW

Powered by Google App Engine
This is Rietveld 408576698