| OLD | NEW | 
|---|
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. | 
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without | 
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are | 
| 4 // met: | 4 // met: | 
| 5 // | 5 // | 
| 6 //     * Redistributions of source code must retain the above copyright | 6 //     * Redistributions of source code must retain the above copyright | 
| 7 //       notice, this list of conditions and the following disclaimer. | 7 //       notice, this list of conditions and the following disclaimer. | 
| 8 //     * Redistributions in binary form must reproduce the above | 8 //     * Redistributions in binary form must reproduce the above | 
| 9 //       copyright notice, this list of conditions and the following | 9 //       copyright notice, this list of conditions and the following | 
| 10 //       disclaimer in the documentation and/or other materials provided | 10 //       disclaimer in the documentation and/or other materials provided | 
| (...skipping 26 matching lines...) Expand all  Loading... | 
| 37 namespace internal { | 37 namespace internal { | 
| 38 | 38 | 
| 39 #define DEFINE_COMPILE(type)                            \ | 39 #define DEFINE_COMPILE(type)                            \ | 
| 40   void L##type::CompileToNative(LCodeGen* generator) {  \ | 40   void L##type::CompileToNative(LCodeGen* generator) {  \ | 
| 41     generator->Do##type(this);                          \ | 41     generator->Do##type(this);                          \ | 
| 42   } | 42   } | 
| 43 LITHIUM_CONCRETE_INSTRUCTION_LIST(DEFINE_COMPILE) | 43 LITHIUM_CONCRETE_INSTRUCTION_LIST(DEFINE_COMPILE) | 
| 44 #undef DEFINE_COMPILE | 44 #undef DEFINE_COMPILE | 
| 45 | 45 | 
| 46 LOsrEntry::LOsrEntry() { | 46 LOsrEntry::LOsrEntry() { | 
| 47   for (int i = 0; i < Register::NumAllocatableRegisters(); ++i) { | 47   for (int i = 0; i < Register::kNumAllocatableRegisters; ++i) { | 
| 48     register_spills_[i] = NULL; | 48     register_spills_[i] = NULL; | 
| 49   } | 49   } | 
| 50   for (int i = 0; i < DoubleRegister::NumAllocatableRegisters(); ++i) { | 50   for (int i = 0; i < DoubleRegister::kNumAllocatableRegisters; ++i) { | 
| 51     double_register_spills_[i] = NULL; | 51     double_register_spills_[i] = NULL; | 
| 52   } | 52   } | 
| 53 } | 53 } | 
| 54 | 54 | 
| 55 | 55 | 
| 56 void LOsrEntry::MarkSpilledRegister(int allocation_index, | 56 void LOsrEntry::MarkSpilledRegister(int allocation_index, | 
| 57                                     LOperand* spill_operand) { | 57                                     LOperand* spill_operand) { | 
| 58   ASSERT(spill_operand->IsStackSlot()); | 58   ASSERT(spill_operand->IsStackSlot()); | 
| 59   ASSERT(register_spills_[allocation_index] == NULL); | 59   ASSERT(register_spills_[allocation_index] == NULL); | 
| 60   register_spills_[allocation_index] = spill_operand; | 60   register_spills_[allocation_index] = spill_operand; | 
| (...skipping 551 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 612   int argument_index_accumulator = 0; | 612   int argument_index_accumulator = 0; | 
| 613   instr->set_environment(CreateEnvironment(hydrogen_env, | 613   instr->set_environment(CreateEnvironment(hydrogen_env, | 
| 614                                            &argument_index_accumulator)); | 614                                            &argument_index_accumulator)); | 
| 615   return instr; | 615   return instr; | 
| 616 } | 616 } | 
| 617 | 617 | 
| 618 | 618 | 
| 619 LInstruction* LChunkBuilder::MarkAsCall(LInstruction* instr, | 619 LInstruction* LChunkBuilder::MarkAsCall(LInstruction* instr, | 
| 620                                         HInstruction* hinstr, | 620                                         HInstruction* hinstr, | 
| 621                                         CanDeoptimize can_deoptimize) { | 621                                         CanDeoptimize can_deoptimize) { | 
| 622   info()->MarkAsNonDeferredCalling(); |  | 
| 623 |  | 
| 624 #ifdef DEBUG | 622 #ifdef DEBUG | 
| 625   instr->VerifyCall(); | 623   instr->VerifyCall(); | 
| 626 #endif | 624 #endif | 
| 627   instr->MarkAsCall(); | 625   instr->MarkAsCall(); | 
| 628   instr = AssignPointerMap(instr); | 626   instr = AssignPointerMap(instr); | 
| 629 | 627 | 
| 630   if (hinstr->HasObservableSideEffects()) { | 628   if (hinstr->HasObservableSideEffects()) { | 
| 631     ASSERT(hinstr->next()->IsSimulate()); | 629     ASSERT(hinstr->next()->IsSimulate()); | 
| 632     HSimulate* sim = HSimulate::cast(hinstr->next()); | 630     HSimulate* sim = HSimulate::cast(hinstr->next()); | 
| 633     ASSERT(instruction_pending_deoptimization_environment_ == NULL); | 631     ASSERT(instruction_pending_deoptimization_environment_ == NULL); | 
| (...skipping 989 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 1623   // All HForceRepresentation instructions should be eliminated in the | 1621   // All HForceRepresentation instructions should be eliminated in the | 
| 1624   // representation change phase of Hydrogen. | 1622   // representation change phase of Hydrogen. | 
| 1625   UNREACHABLE(); | 1623   UNREACHABLE(); | 
| 1626   return NULL; | 1624   return NULL; | 
| 1627 } | 1625 } | 
| 1628 | 1626 | 
| 1629 | 1627 | 
| 1630 LInstruction* LChunkBuilder::DoChange(HChange* instr) { | 1628 LInstruction* LChunkBuilder::DoChange(HChange* instr) { | 
| 1631   Representation from = instr->from(); | 1629   Representation from = instr->from(); | 
| 1632   Representation to = instr->to(); | 1630   Representation to = instr->to(); | 
| 1633   // Only mark conversions that might need to allocate as calling rather than |  | 
| 1634   // all changes. This makes simple, non-allocating conversion not have to force |  | 
| 1635   // building a stack frame. |  | 
| 1636   if (from.IsTagged()) { | 1631   if (from.IsTagged()) { | 
| 1637     if (to.IsDouble()) { | 1632     if (to.IsDouble()) { | 
| 1638       info()->MarkAsDeferredCalling(); |  | 
| 1639       LOperand* value = UseRegister(instr->value()); | 1633       LOperand* value = UseRegister(instr->value()); | 
| 1640       LNumberUntagD* res = new(zone()) LNumberUntagD(value); | 1634       LNumberUntagD* res = new(zone()) LNumberUntagD(value); | 
| 1641       return AssignEnvironment(DefineAsRegister(res)); | 1635       return AssignEnvironment(DefineAsRegister(res)); | 
| 1642     } else { | 1636     } else { | 
| 1643       ASSERT(to.IsInteger32()); | 1637       ASSERT(to.IsInteger32()); | 
| 1644       LOperand* value = UseRegister(instr->value()); | 1638       LOperand* value = UseRegister(instr->value()); | 
| 1645       if (instr->value()->type().IsSmi()) { | 1639       if (instr->value()->type().IsSmi()) { | 
| 1646         return DefineSameAsFirst(new(zone()) LSmiUntag(value, false)); | 1640         return DefineSameAsFirst(new(zone()) LSmiUntag(value, false)); | 
| 1647       } else { | 1641       } else { | 
| 1648         bool truncating = instr->CanTruncateToInt32(); | 1642         bool truncating = instr->CanTruncateToInt32(); | 
| 1649         LOperand* xmm_temp = truncating ? NULL : FixedTemp(xmm1); | 1643         LOperand* xmm_temp = truncating ? NULL : FixedTemp(xmm1); | 
| 1650         LTaggedToI* res = new(zone()) LTaggedToI(value, xmm_temp); | 1644         LTaggedToI* res = new(zone()) LTaggedToI(value, xmm_temp); | 
| 1651         return AssignEnvironment(DefineSameAsFirst(res)); | 1645         return AssignEnvironment(DefineSameAsFirst(res)); | 
| 1652       } | 1646       } | 
| 1653     } | 1647     } | 
| 1654   } else if (from.IsDouble()) { | 1648   } else if (from.IsDouble()) { | 
| 1655     if (to.IsTagged()) { | 1649     if (to.IsTagged()) { | 
| 1656       info()->MarkAsDeferredCalling(); |  | 
| 1657       LOperand* value = UseRegister(instr->value()); | 1650       LOperand* value = UseRegister(instr->value()); | 
| 1658       LOperand* temp = TempRegister(); | 1651       LOperand* temp = TempRegister(); | 
| 1659 | 1652 | 
| 1660       // Make sure that temp and result_temp are different registers. | 1653       // Make sure that temp and result_temp are different registers. | 
| 1661       LUnallocated* result_temp = TempRegister(); | 1654       LUnallocated* result_temp = TempRegister(); | 
| 1662       LNumberTagD* result = new(zone()) LNumberTagD(value, temp); | 1655       LNumberTagD* result = new(zone()) LNumberTagD(value, temp); | 
| 1663       return AssignPointerMap(Define(result, result_temp)); | 1656       return AssignPointerMap(Define(result, result_temp)); | 
| 1664     } else { | 1657     } else { | 
| 1665       ASSERT(to.IsInteger32()); | 1658       ASSERT(to.IsInteger32()); | 
| 1666       LOperand* value = UseRegister(instr->value()); | 1659       LOperand* value = UseRegister(instr->value()); | 
| 1667       return AssignEnvironment(DefineAsRegister(new(zone()) LDoubleToI(value))); | 1660       return AssignEnvironment(DefineAsRegister(new(zone()) LDoubleToI(value))); | 
| 1668     } | 1661     } | 
| 1669   } else if (from.IsInteger32()) { | 1662   } else if (from.IsInteger32()) { | 
| 1670     info()->MarkAsDeferredCalling(); |  | 
| 1671     if (to.IsTagged()) { | 1663     if (to.IsTagged()) { | 
| 1672       HValue* val = instr->value(); | 1664       HValue* val = instr->value(); | 
| 1673       LOperand* value = UseRegister(val); | 1665       LOperand* value = UseRegister(val); | 
| 1674       if (val->CheckFlag(HInstruction::kUint32)) { | 1666       if (val->CheckFlag(HInstruction::kUint32)) { | 
| 1675         LOperand* temp = FixedTemp(xmm1); | 1667         LOperand* temp = FixedTemp(xmm1); | 
| 1676         LNumberTagU* result = new(zone()) LNumberTagU(value, temp); | 1668         LNumberTagU* result = new(zone()) LNumberTagU(value, temp); | 
| 1677         return AssignEnvironment(AssignPointerMap(DefineSameAsFirst(result))); | 1669         return AssignEnvironment(AssignPointerMap(DefineSameAsFirst(result))); | 
| 1678       } else if (val->HasRange() && val->range()->IsInSmiRange()) { | 1670       } else if (val->HasRange() && val->range()->IsInSmiRange()) { | 
| 1679         return DefineSameAsFirst(new(zone()) LSmiTag(value)); | 1671         return DefineSameAsFirst(new(zone()) LSmiTag(value)); | 
| 1680       } else { | 1672       } else { | 
| (...skipping 446 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 2127 | 2119 | 
| 2128 LInstruction* LChunkBuilder::DoOsrEntry(HOsrEntry* instr) { | 2120 LInstruction* LChunkBuilder::DoOsrEntry(HOsrEntry* instr) { | 
| 2129   ASSERT(argument_count_ == 0); | 2121   ASSERT(argument_count_ == 0); | 
| 2130   allocator_->MarkAsOsrEntry(); | 2122   allocator_->MarkAsOsrEntry(); | 
| 2131   current_block_->last_environment()->set_ast_id(instr->ast_id()); | 2123   current_block_->last_environment()->set_ast_id(instr->ast_id()); | 
| 2132   return AssignEnvironment(new(zone()) LOsrEntry); | 2124   return AssignEnvironment(new(zone()) LOsrEntry); | 
| 2133 } | 2125 } | 
| 2134 | 2126 | 
| 2135 | 2127 | 
| 2136 LInstruction* LChunkBuilder::DoParameter(HParameter* instr) { | 2128 LInstruction* LChunkBuilder::DoParameter(HParameter* instr) { | 
| 2137   LParameter* result = new(zone()) LParameter; | 2129   int spill_index = chunk()->GetParameterStackSlot(instr->index()); | 
| 2138   if (info()->IsOptimizing()) { | 2130   return DefineAsSpilled(new(zone()) LParameter, spill_index); | 
| 2139     int spill_index = chunk()->GetParameterStackSlot(instr->index()); |  | 
| 2140     return DefineAsSpilled(result, spill_index); |  | 
| 2141   } else { |  | 
| 2142     ASSERT(info()->IsStub()); |  | 
| 2143     CodeStubInterfaceDescriptor* descriptor = |  | 
| 2144         info()->code_stub()->GetInterfaceDescriptor(info()->isolate()); |  | 
| 2145     Register reg = descriptor->register_params_[instr->index()]; |  | 
| 2146     return DefineFixed(result, reg); |  | 
| 2147   } |  | 
| 2148 } | 2131 } | 
| 2149 | 2132 | 
| 2150 | 2133 | 
| 2151 LInstruction* LChunkBuilder::DoUnknownOSRValue(HUnknownOSRValue* instr) { | 2134 LInstruction* LChunkBuilder::DoUnknownOSRValue(HUnknownOSRValue* instr) { | 
| 2152   int spill_index = chunk()->GetNextSpillIndex(false);  // Not double-width. | 2135   int spill_index = chunk()->GetNextSpillIndex(false);  // Not double-width. | 
| 2153   if (spill_index > LUnallocated::kMaxFixedIndex) { | 2136   if (spill_index > LUnallocated::kMaxFixedIndex) { | 
| 2154     Abort("Too many spill slots needed for OSR"); | 2137     Abort("Too many spill slots needed for OSR"); | 
| 2155     spill_index = 0; | 2138     spill_index = 0; | 
| 2156   } | 2139   } | 
| 2157   return DefineAsSpilled(new(zone()) LUnknownOSRValue, spill_index); | 2140   return DefineAsSpilled(new(zone()) LUnknownOSRValue, spill_index); | 
| (...skipping 75 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 2233     instruction_pending_deoptimization_environment_ = NULL; | 2216     instruction_pending_deoptimization_environment_ = NULL; | 
| 2234     pending_deoptimization_ast_id_ = BailoutId::None(); | 2217     pending_deoptimization_ast_id_ = BailoutId::None(); | 
| 2235     return result; | 2218     return result; | 
| 2236   } | 2219   } | 
| 2237 | 2220 | 
| 2238   return NULL; | 2221   return NULL; | 
| 2239 } | 2222 } | 
| 2240 | 2223 | 
| 2241 | 2224 | 
| 2242 LInstruction* LChunkBuilder::DoStackCheck(HStackCheck* instr) { | 2225 LInstruction* LChunkBuilder::DoStackCheck(HStackCheck* instr) { | 
| 2243   info()->MarkAsDeferredCalling(); |  | 
| 2244   if (instr->is_function_entry()) { | 2226   if (instr->is_function_entry()) { | 
| 2245     return MarkAsCall(new(zone()) LStackCheck, instr); | 2227     return MarkAsCall(new(zone()) LStackCheck, instr); | 
| 2246   } else { | 2228   } else { | 
| 2247     ASSERT(instr->is_backwards_branch()); | 2229     ASSERT(instr->is_backwards_branch()); | 
| 2248     return AssignEnvironment(AssignPointerMap(new(zone()) LStackCheck)); | 2230     return AssignEnvironment(AssignPointerMap(new(zone()) LStackCheck)); | 
| 2249   } | 2231   } | 
| 2250 } | 2232 } | 
| 2251 | 2233 | 
| 2252 | 2234 | 
| 2253 LInstruction* LChunkBuilder::DoEnterInlined(HEnterInlined* instr) { | 2235 LInstruction* LChunkBuilder::DoEnterInlined(HEnterInlined* instr) { | 
| (...skipping 66 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 2320 LInstruction* LChunkBuilder::DoLoadFieldByIndex(HLoadFieldByIndex* instr) { | 2302 LInstruction* LChunkBuilder::DoLoadFieldByIndex(HLoadFieldByIndex* instr) { | 
| 2321   LOperand* object = UseRegister(instr->object()); | 2303   LOperand* object = UseRegister(instr->object()); | 
| 2322   LOperand* index = UseTempRegister(instr->index()); | 2304   LOperand* index = UseTempRegister(instr->index()); | 
| 2323   return DefineSameAsFirst(new(zone()) LLoadFieldByIndex(object, index)); | 2305   return DefineSameAsFirst(new(zone()) LLoadFieldByIndex(object, index)); | 
| 2324 } | 2306 } | 
| 2325 | 2307 | 
| 2326 | 2308 | 
| 2327 } }  // namespace v8::internal | 2309 } }  // namespace v8::internal | 
| 2328 | 2310 | 
| 2329 #endif  // V8_TARGET_ARCH_X64 | 2311 #endif  // V8_TARGET_ARCH_X64 | 
| OLD | NEW | 
|---|