OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 95 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
106 RegisterDependentCodeForEmbeddedMaps(code); | 106 RegisterDependentCodeForEmbeddedMaps(code); |
107 } | 107 } |
108 PopulateDeoptimizationData(code); | 108 PopulateDeoptimizationData(code); |
109 if (!info()->IsStub()) { | 109 if (!info()->IsStub()) { |
110 Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(code); | 110 Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(code); |
111 } | 111 } |
112 info()->CommitDependencies(code); | 112 info()->CommitDependencies(code); |
113 } | 113 } |
114 | 114 |
115 | 115 |
116 void LCodeGen::Abort(BailoutReason reason) { | 116 void LCodeGen::Abort(const char* reason) { |
117 info()->set_bailout_reason(reason); | 117 info()->set_bailout_reason(reason); |
118 status_ = ABORTED; | 118 status_ = ABORTED; |
119 } | 119 } |
120 | 120 |
121 | 121 |
122 void LCodeGen::Comment(const char* format, ...) { | 122 void LCodeGen::Comment(const char* format, ...) { |
123 if (!FLAG_code_comments) return; | 123 if (!FLAG_code_comments) return; |
124 char buffer[4 * KB]; | 124 char buffer[4 * KB]; |
125 StringBuilder builder(buffer, ARRAY_SIZE(buffer)); | 125 StringBuilder builder(buffer, ARRAY_SIZE(buffer)); |
126 va_list arguments; | 126 va_list arguments; |
(...skipping 86 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
213 __ push(Immediate(Smi::FromInt(StackFrame::STUB))); | 213 __ push(Immediate(Smi::FromInt(StackFrame::STUB))); |
214 } else { | 214 } else { |
215 __ push(edi); // Callee's JS function. | 215 __ push(edi); // Callee's JS function. |
216 } | 216 } |
217 } | 217 } |
218 | 218 |
219 if (info()->IsOptimizing() && | 219 if (info()->IsOptimizing() && |
220 dynamic_frame_alignment_ && | 220 dynamic_frame_alignment_ && |
221 FLAG_debug_code) { | 221 FLAG_debug_code) { |
222 __ test(esp, Immediate(kPointerSize)); | 222 __ test(esp, Immediate(kPointerSize)); |
223 __ Assert(zero, kFrameIsExpectedToBeAligned); | 223 __ Assert(zero, "frame is expected to be aligned"); |
224 } | 224 } |
225 | 225 |
226 // Reserve space for the stack slots needed by the code. | 226 // Reserve space for the stack slots needed by the code. |
227 int slots = GetStackSlotCount(); | 227 int slots = GetStackSlotCount(); |
228 ASSERT(slots != 0 || !info()->IsOptimizing()); | 228 ASSERT(slots != 0 || !info()->IsOptimizing()); |
229 if (slots > 0) { | 229 if (slots > 0) { |
230 if (slots == 1) { | 230 if (slots == 1) { |
231 if (dynamic_frame_alignment_) { | 231 if (dynamic_frame_alignment_) { |
232 __ push(edx); | 232 __ push(edx); |
233 } else { | 233 } else { |
(...skipping 641 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
875 void LCodeGen::LoadContextFromDeferred(LOperand* context) { | 875 void LCodeGen::LoadContextFromDeferred(LOperand* context) { |
876 if (context->IsRegister()) { | 876 if (context->IsRegister()) { |
877 if (!ToRegister(context).is(esi)) { | 877 if (!ToRegister(context).is(esi)) { |
878 __ mov(esi, ToRegister(context)); | 878 __ mov(esi, ToRegister(context)); |
879 } | 879 } |
880 } else if (context->IsStackSlot()) { | 880 } else if (context->IsStackSlot()) { |
881 __ mov(esi, ToOperand(context)); | 881 __ mov(esi, ToOperand(context)); |
882 } else if (context->IsConstantOperand()) { | 882 } else if (context->IsConstantOperand()) { |
883 HConstant* constant = | 883 HConstant* constant = |
884 chunk_->LookupConstant(LConstantOperand::cast(context)); | 884 chunk_->LookupConstant(LConstantOperand::cast(context)); |
885 __ LoadObject(esi, Handle<Object>::cast(constant->handle())); | 885 __ LoadHeapObject(esi, Handle<Context>::cast(constant->handle())); |
886 } else { | 886 } else { |
887 UNREACHABLE(); | 887 UNREACHABLE(); |
888 } | 888 } |
889 } | 889 } |
890 | 890 |
891 void LCodeGen::CallRuntimeFromDeferred(Runtime::FunctionId id, | 891 void LCodeGen::CallRuntimeFromDeferred(Runtime::FunctionId id, |
892 int argc, | 892 int argc, |
893 LInstruction* instr, | 893 LInstruction* instr, |
894 LOperand* context) { | 894 LOperand* context) { |
895 LoadContextFromDeferred(context); | 895 LoadContextFromDeferred(context); |
(...skipping 45 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
941 void LCodeGen::DeoptimizeIf(Condition cc, | 941 void LCodeGen::DeoptimizeIf(Condition cc, |
942 LEnvironment* environment, | 942 LEnvironment* environment, |
943 Deoptimizer::BailoutType bailout_type) { | 943 Deoptimizer::BailoutType bailout_type) { |
944 RegisterEnvironmentForDeoptimization(environment, Safepoint::kNoLazyDeopt); | 944 RegisterEnvironmentForDeoptimization(environment, Safepoint::kNoLazyDeopt); |
945 ASSERT(environment->HasBeenRegistered()); | 945 ASSERT(environment->HasBeenRegistered()); |
946 int id = environment->deoptimization_index(); | 946 int id = environment->deoptimization_index(); |
947 ASSERT(info()->IsOptimizing() || info()->IsStub()); | 947 ASSERT(info()->IsOptimizing() || info()->IsStub()); |
948 Address entry = | 948 Address entry = |
949 Deoptimizer::GetDeoptimizationEntry(isolate(), id, bailout_type); | 949 Deoptimizer::GetDeoptimizationEntry(isolate(), id, bailout_type); |
950 if (entry == NULL) { | 950 if (entry == NULL) { |
951 Abort(kBailoutWasNotPrepared); | 951 Abort("bailout was not prepared"); |
952 return; | 952 return; |
953 } | 953 } |
954 | 954 |
955 if (FLAG_deopt_every_n_times != 0 && !info()->IsStub()) { | 955 if (FLAG_deopt_every_n_times != 0 && !info()->IsStub()) { |
956 ExternalReference count = ExternalReference::stress_deopt_count(isolate()); | 956 ExternalReference count = ExternalReference::stress_deopt_count(isolate()); |
957 Label no_deopt; | 957 Label no_deopt; |
958 __ pushfd(); | 958 __ pushfd(); |
959 __ push(eax); | 959 __ push(eax); |
960 __ mov(eax, Operand::StaticVariable(count)); | 960 __ mov(eax, Operand::StaticVariable(count)); |
961 __ sub(eax, Immediate(1)); | 961 __ sub(eax, Immediate(1)); |
(...skipping 710 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1672 } | 1672 } |
1673 | 1673 |
1674 | 1674 |
1675 void LCodeGen::DoBitI(LBitI* instr) { | 1675 void LCodeGen::DoBitI(LBitI* instr) { |
1676 LOperand* left = instr->left(); | 1676 LOperand* left = instr->left(); |
1677 LOperand* right = instr->right(); | 1677 LOperand* right = instr->right(); |
1678 ASSERT(left->Equals(instr->result())); | 1678 ASSERT(left->Equals(instr->result())); |
1679 ASSERT(left->IsRegister()); | 1679 ASSERT(left->IsRegister()); |
1680 | 1680 |
1681 if (right->IsConstantOperand()) { | 1681 if (right->IsConstantOperand()) { |
1682 int32_t right_operand = | 1682 int right_operand = ToRepresentation(LConstantOperand::cast(right), |
1683 ToRepresentation(LConstantOperand::cast(right), | 1683 instr->hydrogen()->representation()); |
1684 instr->hydrogen()->representation()); | |
1685 switch (instr->op()) { | 1684 switch (instr->op()) { |
1686 case Token::BIT_AND: | 1685 case Token::BIT_AND: |
1687 __ and_(ToRegister(left), right_operand); | 1686 __ and_(ToRegister(left), right_operand); |
1688 break; | 1687 break; |
1689 case Token::BIT_OR: | 1688 case Token::BIT_OR: |
1690 __ or_(ToRegister(left), right_operand); | 1689 __ or_(ToRegister(left), right_operand); |
1691 break; | 1690 break; |
1692 case Token::BIT_XOR: | 1691 case Token::BIT_XOR: |
1693 if (right_operand == int32_t(~0)) { | 1692 __ xor_(ToRegister(left), right_operand); |
1694 __ not_(ToRegister(left)); | |
1695 } else { | |
1696 __ xor_(ToRegister(left), right_operand); | |
1697 } | |
1698 break; | 1693 break; |
1699 default: | 1694 default: |
1700 UNREACHABLE(); | 1695 UNREACHABLE(); |
1701 break; | 1696 break; |
1702 } | 1697 } |
1703 } else { | 1698 } else { |
1704 switch (instr->op()) { | 1699 switch (instr->op()) { |
1705 case Token::BIT_AND: | 1700 case Token::BIT_AND: |
1706 __ and_(ToRegister(left), ToOperand(right)); | 1701 __ and_(ToRegister(left), ToOperand(right)); |
1707 break; | 1702 break; |
(...skipping 266 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1974 if (FLAG_debug_code) { | 1969 if (FLAG_debug_code) { |
1975 __ push(value); | 1970 __ push(value); |
1976 __ mov(value, FieldOperand(string, HeapObject::kMapOffset)); | 1971 __ mov(value, FieldOperand(string, HeapObject::kMapOffset)); |
1977 __ movzx_b(value, FieldOperand(value, Map::kInstanceTypeOffset)); | 1972 __ movzx_b(value, FieldOperand(value, Map::kInstanceTypeOffset)); |
1978 | 1973 |
1979 __ and_(value, Immediate(kStringRepresentationMask | kStringEncodingMask)); | 1974 __ and_(value, Immediate(kStringRepresentationMask | kStringEncodingMask)); |
1980 static const uint32_t one_byte_seq_type = kSeqStringTag | kOneByteStringTag; | 1975 static const uint32_t one_byte_seq_type = kSeqStringTag | kOneByteStringTag; |
1981 static const uint32_t two_byte_seq_type = kSeqStringTag | kTwoByteStringTag; | 1976 static const uint32_t two_byte_seq_type = kSeqStringTag | kTwoByteStringTag; |
1982 __ cmp(value, Immediate(encoding == String::ONE_BYTE_ENCODING | 1977 __ cmp(value, Immediate(encoding == String::ONE_BYTE_ENCODING |
1983 ? one_byte_seq_type : two_byte_seq_type)); | 1978 ? one_byte_seq_type : two_byte_seq_type)); |
1984 __ Check(equal, kUnexpectedStringType); | 1979 __ Check(equal, "Unexpected string type"); |
1985 __ pop(value); | 1980 __ pop(value); |
1986 } | 1981 } |
1987 | 1982 |
1988 if (encoding == String::ONE_BYTE_ENCODING) { | 1983 if (encoding == String::ONE_BYTE_ENCODING) { |
1989 __ mov_b(FieldOperand(string, index, times_1, SeqString::kHeaderSize), | 1984 __ mov_b(FieldOperand(string, index, times_1, SeqString::kHeaderSize), |
1990 value); | 1985 value); |
1991 } else { | 1986 } else { |
1992 __ mov_w(FieldOperand(string, index, times_2, SeqString::kHeaderSize), | 1987 __ mov_w(FieldOperand(string, index, times_2, SeqString::kHeaderSize), |
1993 value); | 1988 value); |
1994 } | 1989 } |
1995 } | 1990 } |
1996 | 1991 |
1997 | 1992 |
| 1993 void LCodeGen::DoBitNotI(LBitNotI* instr) { |
| 1994 LOperand* input = instr->value(); |
| 1995 ASSERT(input->Equals(instr->result())); |
| 1996 __ not_(ToRegister(input)); |
| 1997 } |
| 1998 |
| 1999 |
1998 void LCodeGen::DoThrow(LThrow* instr) { | 2000 void LCodeGen::DoThrow(LThrow* instr) { |
1999 __ push(ToOperand(instr->value())); | 2001 __ push(ToOperand(instr->value())); |
2000 ASSERT(ToRegister(instr->context()).is(esi)); | 2002 ASSERT(ToRegister(instr->context()).is(esi)); |
2001 CallRuntime(Runtime::kThrow, 1, instr); | 2003 CallRuntime(Runtime::kThrow, 1, instr); |
2002 | 2004 |
2003 if (FLAG_debug_code) { | 2005 if (FLAG_debug_code) { |
2004 Comment("Unreachable code."); | 2006 Comment("Unreachable code."); |
2005 __ int3(); | 2007 __ int3(); |
2006 } | 2008 } |
2007 } | 2009 } |
(...skipping 846 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2854 | 2856 |
2855 void LCodeGen::EmitReturn(LReturn* instr, bool dynamic_frame_alignment) { | 2857 void LCodeGen::EmitReturn(LReturn* instr, bool dynamic_frame_alignment) { |
2856 int extra_value_count = dynamic_frame_alignment ? 2 : 1; | 2858 int extra_value_count = dynamic_frame_alignment ? 2 : 1; |
2857 | 2859 |
2858 if (instr->has_constant_parameter_count()) { | 2860 if (instr->has_constant_parameter_count()) { |
2859 int parameter_count = ToInteger32(instr->constant_parameter_count()); | 2861 int parameter_count = ToInteger32(instr->constant_parameter_count()); |
2860 if (dynamic_frame_alignment && FLAG_debug_code) { | 2862 if (dynamic_frame_alignment && FLAG_debug_code) { |
2861 __ cmp(Operand(esp, | 2863 __ cmp(Operand(esp, |
2862 (parameter_count + extra_value_count) * kPointerSize), | 2864 (parameter_count + extra_value_count) * kPointerSize), |
2863 Immediate(kAlignmentZapValue)); | 2865 Immediate(kAlignmentZapValue)); |
2864 __ Assert(equal, kExpectedAlignmentMarker); | 2866 __ Assert(equal, "expected alignment marker"); |
2865 } | 2867 } |
2866 __ Ret((parameter_count + extra_value_count) * kPointerSize, ecx); | 2868 __ Ret((parameter_count + extra_value_count) * kPointerSize, ecx); |
2867 } else { | 2869 } else { |
2868 Register reg = ToRegister(instr->parameter_count()); | 2870 Register reg = ToRegister(instr->parameter_count()); |
2869 // The argument count parameter is a smi | 2871 // The argument count parameter is a smi |
2870 __ SmiUntag(reg); | 2872 __ SmiUntag(reg); |
2871 Register return_addr_reg = reg.is(ecx) ? ebx : ecx; | 2873 Register return_addr_reg = reg.is(ecx) ? ebx : ecx; |
2872 if (dynamic_frame_alignment && FLAG_debug_code) { | 2874 if (dynamic_frame_alignment && FLAG_debug_code) { |
2873 ASSERT(extra_value_count == 2); | 2875 ASSERT(extra_value_count == 2); |
2874 __ cmp(Operand(esp, reg, times_pointer_size, | 2876 __ cmp(Operand(esp, reg, times_pointer_size, |
2875 extra_value_count * kPointerSize), | 2877 extra_value_count * kPointerSize), |
2876 Immediate(kAlignmentZapValue)); | 2878 Immediate(kAlignmentZapValue)); |
2877 __ Assert(equal, kExpectedAlignmentMarker); | 2879 __ Assert(equal, "expected alignment marker"); |
2878 } | 2880 } |
2879 | 2881 |
2880 // emit code to restore stack based on instr->parameter_count() | 2882 // emit code to restore stack based on instr->parameter_count() |
2881 __ pop(return_addr_reg); // save return address | 2883 __ pop(return_addr_reg); // save return address |
2882 if (dynamic_frame_alignment) { | 2884 if (dynamic_frame_alignment) { |
2883 __ inc(reg); // 1 more for alignment | 2885 __ inc(reg); // 1 more for alignment |
2884 } | 2886 } |
2885 __ shl(reg, kPointerSizeLog2); | 2887 __ shl(reg, kPointerSizeLog2); |
2886 __ add(esp, reg); | 2888 __ add(esp, reg); |
2887 __ jmp(return_addr_reg); | 2889 __ jmp(return_addr_reg); |
(...skipping 550 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3438 Representation key_representation, | 3440 Representation key_representation, |
3439 ElementsKind elements_kind, | 3441 ElementsKind elements_kind, |
3440 uint32_t offset, | 3442 uint32_t offset, |
3441 uint32_t additional_index) { | 3443 uint32_t additional_index) { |
3442 Register elements_pointer_reg = ToRegister(elements_pointer); | 3444 Register elements_pointer_reg = ToRegister(elements_pointer); |
3443 int element_shift_size = ElementsKindToShiftSize(elements_kind); | 3445 int element_shift_size = ElementsKindToShiftSize(elements_kind); |
3444 int shift_size = element_shift_size; | 3446 int shift_size = element_shift_size; |
3445 if (key->IsConstantOperand()) { | 3447 if (key->IsConstantOperand()) { |
3446 int constant_value = ToInteger32(LConstantOperand::cast(key)); | 3448 int constant_value = ToInteger32(LConstantOperand::cast(key)); |
3447 if (constant_value & 0xF0000000) { | 3449 if (constant_value & 0xF0000000) { |
3448 Abort(kArrayIndexConstantValueTooBig); | 3450 Abort("array index constant value too big"); |
3449 } | 3451 } |
3450 return Operand(elements_pointer_reg, | 3452 return Operand(elements_pointer_reg, |
3451 ((constant_value + additional_index) << shift_size) | 3453 ((constant_value + additional_index) << shift_size) |
3452 + offset); | 3454 + offset); |
3453 } else { | 3455 } else { |
3454 // Take the tag bit into account while computing the shift size. | 3456 // Take the tag bit into account while computing the shift size. |
3455 if (key_representation.IsSmi() && (shift_size >= 1)) { | 3457 if (key_representation.IsSmi() && (shift_size >= 1)) { |
3456 shift_size -= kSmiTagSize; | 3458 shift_size -= kSmiTagSize; |
3457 } | 3459 } |
3458 ScaleFactor scale_factor = static_cast<ScaleFactor>(shift_size); | 3460 ScaleFactor scale_factor = static_cast<ScaleFactor>(shift_size); |
(...skipping 353 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3812 ASSERT(instr->value()->Equals(instr->result())); | 3814 ASSERT(instr->value()->Equals(instr->result())); |
3813 Representation r = instr->hydrogen()->value()->representation(); | 3815 Representation r = instr->hydrogen()->value()->representation(); |
3814 | 3816 |
3815 CpuFeatureScope scope(masm(), SSE2); | 3817 CpuFeatureScope scope(masm(), SSE2); |
3816 if (r.IsDouble()) { | 3818 if (r.IsDouble()) { |
3817 XMMRegister scratch = xmm0; | 3819 XMMRegister scratch = xmm0; |
3818 XMMRegister input_reg = ToDoubleRegister(instr->value()); | 3820 XMMRegister input_reg = ToDoubleRegister(instr->value()); |
3819 __ xorps(scratch, scratch); | 3821 __ xorps(scratch, scratch); |
3820 __ subsd(scratch, input_reg); | 3822 __ subsd(scratch, input_reg); |
3821 __ pand(input_reg, scratch); | 3823 __ pand(input_reg, scratch); |
3822 } else if (r.IsSmiOrInteger32()) { | 3824 } else if (r.IsInteger32()) { |
3823 EmitIntegerMathAbs(instr); | 3825 EmitIntegerMathAbs(instr); |
3824 } else { // Tagged case. | 3826 } else { // Tagged case. |
3825 DeferredMathAbsTaggedHeapNumber* deferred = | 3827 DeferredMathAbsTaggedHeapNumber* deferred = |
3826 new(zone()) DeferredMathAbsTaggedHeapNumber(this, instr); | 3828 new(zone()) DeferredMathAbsTaggedHeapNumber(this, instr); |
3827 Register input_reg = ToRegister(instr->value()); | 3829 Register input_reg = ToRegister(instr->value()); |
3828 // Smi check. | 3830 // Smi check. |
3829 __ JumpIfNotSmi(input_reg, deferred->entry()); | 3831 __ JumpIfNotSmi(input_reg, deferred->entry()); |
3830 EmitIntegerMathAbs(instr); | 3832 EmitIntegerMathAbs(instr); |
3831 __ bind(deferred->exit()); | 3833 __ bind(deferred->exit()); |
3832 } | 3834 } |
(...skipping 1951 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5784 Handle<Cell> cell = isolate()->factory()->NewCell(target); | 5786 Handle<Cell> cell = isolate()->factory()->NewCell(target); |
5785 __ cmp(reg, Operand::ForCell(cell)); | 5787 __ cmp(reg, Operand::ForCell(cell)); |
5786 } else { | 5788 } else { |
5787 Operand operand = ToOperand(instr->value()); | 5789 Operand operand = ToOperand(instr->value()); |
5788 __ cmp(operand, target); | 5790 __ cmp(operand, target); |
5789 } | 5791 } |
5790 DeoptimizeIf(not_equal, instr->environment()); | 5792 DeoptimizeIf(not_equal, instr->environment()); |
5791 } | 5793 } |
5792 | 5794 |
5793 | 5795 |
5794 void LCodeGen::DoDeferredInstanceMigration(LCheckMaps* instr, Register object) { | 5796 void LCodeGen::DoCheckMapCommon(Register reg, |
5795 { | 5797 Handle<Map> map, |
5796 PushSafepointRegistersScope scope(this); | 5798 LInstruction* instr) { |
5797 __ push(object); | 5799 Label success; |
5798 __ xor_(esi, esi); | 5800 __ CompareMap(reg, map, &success); |
5799 __ CallRuntimeSaveDoubles(Runtime::kMigrateInstance); | 5801 DeoptimizeIf(not_equal, instr->environment()); |
5800 RecordSafepointWithRegisters( | 5802 __ bind(&success); |
5801 instr->pointer_map(), 1, Safepoint::kNoLazyDeopt); | |
5802 | |
5803 __ test(eax, Immediate(kSmiTagMask)); | |
5804 } | |
5805 DeoptimizeIf(zero, instr->environment()); | |
5806 } | 5803 } |
5807 | 5804 |
5808 | 5805 |
5809 void LCodeGen::DoCheckMaps(LCheckMaps* instr) { | 5806 void LCodeGen::DoCheckMaps(LCheckMaps* instr) { |
5810 class DeferredCheckMaps: public LDeferredCode { | |
5811 public: | |
5812 DeferredCheckMaps(LCodeGen* codegen, LCheckMaps* instr, Register object) | |
5813 : LDeferredCode(codegen), instr_(instr), object_(object) { | |
5814 SetExit(check_maps()); | |
5815 } | |
5816 virtual void Generate() { | |
5817 codegen()->DoDeferredInstanceMigration(instr_, object_); | |
5818 } | |
5819 Label* check_maps() { return &check_maps_; } | |
5820 virtual LInstruction* instr() { return instr_; } | |
5821 private: | |
5822 LCheckMaps* instr_; | |
5823 Label check_maps_; | |
5824 Register object_; | |
5825 }; | |
5826 | |
5827 if (instr->hydrogen()->CanOmitMapChecks()) return; | 5807 if (instr->hydrogen()->CanOmitMapChecks()) return; |
5828 | |
5829 LOperand* input = instr->value(); | 5808 LOperand* input = instr->value(); |
5830 ASSERT(input->IsRegister()); | 5809 ASSERT(input->IsRegister()); |
5831 Register reg = ToRegister(input); | 5810 Register reg = ToRegister(input); |
5832 | 5811 |
| 5812 Label success; |
5833 SmallMapList* map_set = instr->hydrogen()->map_set(); | 5813 SmallMapList* map_set = instr->hydrogen()->map_set(); |
5834 | |
5835 DeferredCheckMaps* deferred = NULL; | |
5836 if (instr->hydrogen()->has_migration_target()) { | |
5837 deferred = new(zone()) DeferredCheckMaps(this, instr, reg); | |
5838 __ bind(deferred->check_maps()); | |
5839 } | |
5840 | |
5841 Label success; | |
5842 for (int i = 0; i < map_set->length() - 1; i++) { | 5814 for (int i = 0; i < map_set->length() - 1; i++) { |
5843 Handle<Map> map = map_set->at(i); | 5815 Handle<Map> map = map_set->at(i); |
5844 __ CompareMap(reg, map, &success); | 5816 __ CompareMap(reg, map, &success); |
5845 __ j(equal, &success); | 5817 __ j(equal, &success); |
5846 } | 5818 } |
5847 | |
5848 Handle<Map> map = map_set->last(); | 5819 Handle<Map> map = map_set->last(); |
5849 __ CompareMap(reg, map, &success); | 5820 DoCheckMapCommon(reg, map, instr); |
5850 if (instr->hydrogen()->has_migration_target()) { | |
5851 __ j(not_equal, deferred->entry()); | |
5852 } else { | |
5853 DeoptimizeIf(not_equal, instr->environment()); | |
5854 } | |
5855 | |
5856 __ bind(&success); | 5821 __ bind(&success); |
5857 } | 5822 } |
5858 | 5823 |
5859 | 5824 |
5860 void LCodeGen::DoClampDToUint8(LClampDToUint8* instr) { | 5825 void LCodeGen::DoClampDToUint8(LClampDToUint8* instr) { |
5861 CpuFeatureScope scope(masm(), SSE2); | 5826 CpuFeatureScope scope(masm(), SSE2); |
5862 XMMRegister value_reg = ToDoubleRegister(instr->unclamped()); | 5827 XMMRegister value_reg = ToDoubleRegister(instr->unclamped()); |
5863 Register result_reg = ToRegister(instr->result()); | 5828 Register result_reg = ToRegister(instr->result()); |
5864 __ ClampDoubleToUint8(value_reg, xmm0, result_reg); | 5829 __ ClampDoubleToUint8(value_reg, xmm0, result_reg); |
5865 } | 5830 } |
(...skipping 156 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6022 __ bind(&is_smi); | 5987 __ bind(&is_smi); |
6023 if (!input_reg.is(result_reg)) { | 5988 if (!input_reg.is(result_reg)) { |
6024 __ mov(result_reg, input_reg); | 5989 __ mov(result_reg, input_reg); |
6025 } | 5990 } |
6026 __ SmiUntag(result_reg); | 5991 __ SmiUntag(result_reg); |
6027 __ ClampUint8(result_reg); | 5992 __ ClampUint8(result_reg); |
6028 __ bind(&done); | 5993 __ bind(&done); |
6029 } | 5994 } |
6030 | 5995 |
6031 | 5996 |
| 5997 void LCodeGen::DoCheckPrototypeMaps(LCheckPrototypeMaps* instr) { |
| 5998 if (instr->hydrogen()->CanOmitPrototypeChecks()) return; |
| 5999 Register reg = ToRegister(instr->temp()); |
| 6000 |
| 6001 ZoneList<Handle<JSObject> >* prototypes = instr->prototypes(); |
| 6002 ZoneList<Handle<Map> >* maps = instr->maps(); |
| 6003 |
| 6004 ASSERT(prototypes->length() == maps->length()); |
| 6005 |
| 6006 for (int i = 0; i < prototypes->length(); i++) { |
| 6007 __ LoadHeapObject(reg, prototypes->at(i)); |
| 6008 DoCheckMapCommon(reg, maps->at(i), instr); |
| 6009 } |
| 6010 } |
| 6011 |
| 6012 |
6032 void LCodeGen::DoAllocate(LAllocate* instr) { | 6013 void LCodeGen::DoAllocate(LAllocate* instr) { |
6033 class DeferredAllocate: public LDeferredCode { | 6014 class DeferredAllocate: public LDeferredCode { |
6034 public: | 6015 public: |
6035 DeferredAllocate(LCodeGen* codegen, LAllocate* instr) | 6016 DeferredAllocate(LCodeGen* codegen, LAllocate* instr) |
6036 : LDeferredCode(codegen), instr_(instr) { } | 6017 : LDeferredCode(codegen), instr_(instr) { } |
6037 virtual void Generate() { codegen()->DoDeferredAllocate(instr_); } | 6018 virtual void Generate() { codegen()->DoDeferredAllocate(instr_); } |
6038 virtual LInstruction* instr() { return instr_; } | 6019 virtual LInstruction* instr() { return instr_; } |
6039 private: | 6020 private: |
6040 LAllocate* instr_; | 6021 LAllocate* instr_; |
6041 }; | 6022 }; |
(...skipping 488 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6530 FixedArray::kHeaderSize - kPointerSize)); | 6511 FixedArray::kHeaderSize - kPointerSize)); |
6531 __ bind(&done); | 6512 __ bind(&done); |
6532 } | 6513 } |
6533 | 6514 |
6534 | 6515 |
6535 #undef __ | 6516 #undef __ |
6536 | 6517 |
6537 } } // namespace v8::internal | 6518 } } // namespace v8::internal |
6538 | 6519 |
6539 #endif // V8_TARGET_ARCH_IA32 | 6520 #endif // V8_TARGET_ARCH_IA32 |
OLD | NEW |