| OLD | NEW |
| 1 // Copyright 2013 the V8 project authors. All rights reserved. | 1 // Copyright 2013 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 679 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 690 | 690 |
| 691 void LCodeGen::DeoptimizeIf(Condition cc, | 691 void LCodeGen::DeoptimizeIf(Condition cc, |
| 692 LEnvironment* environment) { | 692 LEnvironment* environment) { |
| 693 Deoptimizer::BailoutType bailout_type = info()->IsStub() | 693 Deoptimizer::BailoutType bailout_type = info()->IsStub() |
| 694 ? Deoptimizer::LAZY | 694 ? Deoptimizer::LAZY |
| 695 : Deoptimizer::EAGER; | 695 : Deoptimizer::EAGER; |
| 696 DeoptimizeIf(cc, environment, bailout_type); | 696 DeoptimizeIf(cc, environment, bailout_type); |
| 697 } | 697 } |
| 698 | 698 |
| 699 | 699 |
| 700 void LCodeGen::SoftDeoptimize(LEnvironment* environment) { | |
| 701 ASSERT(!info()->IsStub()); | |
| 702 DeoptimizeIf(no_condition, environment, Deoptimizer::SOFT); | |
| 703 } | |
| 704 | |
| 705 | |
| 706 void LCodeGen::RegisterDependentCodeForEmbeddedMaps(Handle<Code> code) { | 700 void LCodeGen::RegisterDependentCodeForEmbeddedMaps(Handle<Code> code) { |
| 707 ZoneList<Handle<Map> > maps(1, zone()); | 701 ZoneList<Handle<Map> > maps(1, zone()); |
| 708 int mode_mask = RelocInfo::ModeMask(RelocInfo::EMBEDDED_OBJECT); | 702 int mode_mask = RelocInfo::ModeMask(RelocInfo::EMBEDDED_OBJECT); |
| 709 for (RelocIterator it(*code, mode_mask); !it.done(); it.next()) { | 703 for (RelocIterator it(*code, mode_mask); !it.done(); it.next()) { |
| 710 RelocInfo::Mode mode = it.rinfo()->rmode(); | 704 RelocInfo::Mode mode = it.rinfo()->rmode(); |
| 711 if (mode == RelocInfo::EMBEDDED_OBJECT && | 705 if (mode == RelocInfo::EMBEDDED_OBJECT && |
| 712 it.rinfo()->target_object()->IsMap()) { | 706 it.rinfo()->target_object()->IsMap()) { |
| 713 Handle<Map> map(Map::cast(it.rinfo()->target_object())); | 707 Handle<Map> map(Map::cast(it.rinfo()->target_object())); |
| 714 if (map->CanTransition()) { | 708 if (map->CanTransition()) { |
| 715 maps.Add(map, zone()); | 709 maps.Add(map, zone()); |
| (...skipping 4641 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5357 EnsureSpaceForLazyDeopt(Deoptimizer::patch_size()); | 5351 EnsureSpaceForLazyDeopt(Deoptimizer::patch_size()); |
| 5358 last_lazy_deopt_pc_ = masm()->pc_offset(); | 5352 last_lazy_deopt_pc_ = masm()->pc_offset(); |
| 5359 ASSERT(instr->HasEnvironment()); | 5353 ASSERT(instr->HasEnvironment()); |
| 5360 LEnvironment* env = instr->environment(); | 5354 LEnvironment* env = instr->environment(); |
| 5361 RegisterEnvironmentForDeoptimization(env, Safepoint::kLazyDeopt); | 5355 RegisterEnvironmentForDeoptimization(env, Safepoint::kLazyDeopt); |
| 5362 safepoints_.RecordLazyDeoptimizationIndex(env->deoptimization_index()); | 5356 safepoints_.RecordLazyDeoptimizationIndex(env->deoptimization_index()); |
| 5363 } | 5357 } |
| 5364 | 5358 |
| 5365 | 5359 |
| 5366 void LCodeGen::DoDeoptimize(LDeoptimize* instr) { | 5360 void LCodeGen::DoDeoptimize(LDeoptimize* instr) { |
| 5367 if (instr->hydrogen_value()->IsSoftDeoptimize()) { | 5361 Deoptimizer::BailoutType type = instr->hydrogen()->type(); |
| 5368 SoftDeoptimize(instr->environment()); | 5362 // TODO(danno): Stubs expect all deopts to be lazy for historical reasons (the |
| 5369 } else { | 5363 // needed return address), even though the implementation of LAZY and EAGER is |
| 5370 DeoptimizeIf(no_condition, instr->environment()); | 5364 // now identical. When LAZY is eventually completely folded into EAGER, remove |
| 5365 // the special case below. |
| 5366 if (info()->IsStub() && type == Deoptimizer::EAGER) { |
| 5367 type = Deoptimizer::LAZY; |
| 5371 } | 5368 } |
| 5369 DeoptimizeIf(no_condition, instr->environment(), type); |
| 5372 } | 5370 } |
| 5373 | 5371 |
| 5374 | 5372 |
| 5375 void LCodeGen::DoDummyUse(LDummyUse* instr) { | 5373 void LCodeGen::DoDummyUse(LDummyUse* instr) { |
| 5376 // Nothing to see here, move on! | 5374 // Nothing to see here, move on! |
| 5377 } | 5375 } |
| 5378 | 5376 |
| 5379 | 5377 |
| 5380 void LCodeGen::DoDeferredStackCheck(LStackCheck* instr) { | 5378 void LCodeGen::DoDeferredStackCheck(LStackCheck* instr) { |
| 5381 PushSafepointRegistersScope scope(this); | 5379 PushSafepointRegistersScope scope(this); |
| (...skipping 156 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5538 FixedArray::kHeaderSize - kPointerSize)); | 5536 FixedArray::kHeaderSize - kPointerSize)); |
| 5539 __ bind(&done); | 5537 __ bind(&done); |
| 5540 } | 5538 } |
| 5541 | 5539 |
| 5542 | 5540 |
| 5543 #undef __ | 5541 #undef __ |
| 5544 | 5542 |
| 5545 } } // namespace v8::internal | 5543 } } // namespace v8::internal |
| 5546 | 5544 |
| 5547 #endif // V8_TARGET_ARCH_X64 | 5545 #endif // V8_TARGET_ARCH_X64 |
| OLD | NEW |