Index: src/mips64/code-stubs-mips64.cc |
diff --git a/src/mips64/code-stubs-mips64.cc b/src/mips64/code-stubs-mips64.cc |
index 599ca16232da77f822ddea9aa05a4266ea44a845..970792aafa9baff7c9f2a4b2d8aedd2309d2fe61 100644 |
--- a/src/mips64/code-stubs-mips64.cc |
+++ b/src/mips64/code-stubs-mips64.cc |
@@ -357,7 +357,7 @@ void HydrogenCodeStub::GenerateLightweightMiss(MacroAssembler* masm) { |
{ |
// Call the runtime system in a fresh internal frame. |
FrameScope scope(masm, StackFrame::INTERNAL); |
- ASSERT((param_count == 0) || |
+ DCHECK((param_count == 0) || |
a0.is(descriptor->GetEnvironmentParameterRegister(param_count - 1))); |
// Push arguments, adjust sp. |
__ Dsubu(sp, sp, Operand(param_count * kPointerSize)); |
@@ -645,7 +645,7 @@ void WriteInt32ToHeapNumberStub::Generate(MacroAssembler* masm) { |
// but it just ends up combining harmlessly with the last digit of the |
// exponent that happens to be 1. The sign bit is 0 so we shift 10 to get |
// the most significant 1 to hit the last bit of the 12 bit sign and exponent. |
- ASSERT(((1 << HeapNumber::kExponentShift) & non_smi_exponent) != 0); |
+ DCHECK(((1 << HeapNumber::kExponentShift) & non_smi_exponent) != 0); |
const int shift_distance = HeapNumber::kNonMantissaBitsInTopWord - 2; |
__ srl(at, the_int_, shift_distance); |
__ or_(scratch_, scratch_, at); |
@@ -706,7 +706,7 @@ static void EmitIdenticalObjectComparison(MacroAssembler* masm, |
__ Branch(&return_equal, ne, t0, Operand(ODDBALL_TYPE)); |
__ LoadRoot(a6, Heap::kUndefinedValueRootIndex); |
__ Branch(&return_equal, ne, a0, Operand(a6)); |
- ASSERT(is_int16(GREATER) && is_int16(LESS)); |
+ DCHECK(is_int16(GREATER) && is_int16(LESS)); |
__ Ret(USE_DELAY_SLOT); |
if (cc == le) { |
// undefined <= undefined should fail. |
@@ -720,7 +720,7 @@ static void EmitIdenticalObjectComparison(MacroAssembler* masm, |
} |
__ bind(&return_equal); |
- ASSERT(is_int16(GREATER) && is_int16(LESS)); |
+ DCHECK(is_int16(GREATER) && is_int16(LESS)); |
__ Ret(USE_DELAY_SLOT); |
if (cc == less) { |
__ li(v0, Operand(GREATER)); // Things aren't less than themselves. |
@@ -758,7 +758,7 @@ static void EmitIdenticalObjectComparison(MacroAssembler* masm, |
if (cc != eq) { |
// All-zero means Infinity means equal. |
__ Ret(eq, v0, Operand(zero_reg)); |
- ASSERT(is_int16(GREATER) && is_int16(LESS)); |
+ DCHECK(is_int16(GREATER) && is_int16(LESS)); |
__ Ret(USE_DELAY_SLOT); |
if (cc == le) { |
__ li(v0, Operand(GREATER)); // NaN <= NaN should fail. |
@@ -779,7 +779,7 @@ static void EmitSmiNonsmiComparison(MacroAssembler* masm, |
Label* both_loaded_as_doubles, |
Label* slow, |
bool strict) { |
- ASSERT((lhs.is(a0) && rhs.is(a1)) || |
+ DCHECK((lhs.is(a0) && rhs.is(a1)) || |
(lhs.is(a1) && rhs.is(a0))); |
Label lhs_is_smi; |
@@ -896,7 +896,7 @@ static void EmitCheckForInternalizedStringsOrObjects(MacroAssembler* masm, |
Register rhs, |
Label* possible_strings, |
Label* not_both_strings) { |
- ASSERT((lhs.is(a0) && rhs.is(a1)) || |
+ DCHECK((lhs.is(a0) && rhs.is(a1)) || |
(lhs.is(a1) && rhs.is(a0))); |
// a2 is object type of rhs. |
@@ -988,7 +988,7 @@ void ICCompareStub::GenerateGeneric(MacroAssembler* masm) { |
// If either is a Smi (we know that not both are), then they can only |
// be strictly equal if the other is a HeapNumber. |
STATIC_ASSERT(kSmiTag == 0); |
- ASSERT_EQ(0, Smi::FromInt(0)); |
+ DCHECK_EQ(0, Smi::FromInt(0)); |
__ And(a6, lhs, Operand(rhs)); |
__ JumpIfNotSmi(a6, ¬_smis, a4); |
// One operand is a smi. EmitSmiNonsmiComparison generates code that can: |
@@ -1044,7 +1044,7 @@ void ICCompareStub::GenerateGeneric(MacroAssembler* masm) { |
__ bind(&nan); |
// NaN comparisons always fail. |
// Load whatever we need in v0 to make the comparison fail. |
- ASSERT(is_int16(GREATER) && is_int16(LESS)); |
+ DCHECK(is_int16(GREATER) && is_int16(LESS)); |
__ Ret(USE_DELAY_SLOT); |
if (cc == lt || cc == le) { |
__ li(v0, Operand(GREATER)); |
@@ -1126,7 +1126,7 @@ void ICCompareStub::GenerateGeneric(MacroAssembler* masm) { |
if (cc == lt || cc == le) { |
ncr = GREATER; |
} else { |
- ASSERT(cc == gt || cc == ge); // Remaining cases. |
+ DCHECK(cc == gt || cc == ge); // Remaining cases. |
ncr = LESS; |
} |
__ li(a0, Operand(Smi::FromInt(ncr))); |
@@ -1377,7 +1377,7 @@ void MathPowStub::Generate(MacroAssembler* masm) { |
heapnumber, scratch, scratch2, heapnumbermap, &call_runtime); |
__ sdc1(double_result, |
FieldMemOperand(heapnumber, HeapNumber::kValueOffset)); |
- ASSERT(heapnumber.is(v0)); |
+ DCHECK(heapnumber.is(v0)); |
__ IncrementCounter(counters->math_pow(), 1, scratch, scratch2); |
__ DropAndRet(2); |
} else { |
@@ -1524,7 +1524,7 @@ void CEntryStub::Generate(MacroAssembler* masm) { |
// Set up sp in the delay slot. |
masm->daddiu(sp, sp, -kCArgsSlotsSize); |
// Make sure the stored 'ra' points to this position. |
- ASSERT_EQ(kNumInstructionsToJump, |
+ DCHECK_EQ(kNumInstructionsToJump, |
masm->InstructionsGeneratedSince(&find_ra)); |
} |
@@ -1779,9 +1779,9 @@ void JSEntryStub::GenerateBody(MacroAssembler* masm, bool is_construct) { |
// in the safepoint slot for register a4. |
void InstanceofStub::Generate(MacroAssembler* masm) { |
// Call site inlining and patching implies arguments in registers. |
- ASSERT(HasArgsInRegisters() || !HasCallSiteInlineCheck()); |
+ DCHECK(HasArgsInRegisters() || !HasCallSiteInlineCheck()); |
// ReturnTrueFalse is only implemented for inlined call sites. |
- ASSERT(!ReturnTrueFalseObject() || HasCallSiteInlineCheck()); |
+ DCHECK(!ReturnTrueFalseObject() || HasCallSiteInlineCheck()); |
// Fixed register usage throughout the stub: |
const Register object = a0; // Object (lhs). |
@@ -1831,7 +1831,7 @@ void InstanceofStub::Generate(MacroAssembler* masm) { |
__ StoreRoot(function, Heap::kInstanceofCacheFunctionRootIndex); |
__ StoreRoot(map, Heap::kInstanceofCacheMapRootIndex); |
} else { |
- ASSERT(HasArgsInRegisters()); |
+ DCHECK(HasArgsInRegisters()); |
// Patch the (relocated) inlined map check. |
// The offset was stored in a4 safepoint slot. |
@@ -1861,7 +1861,7 @@ void InstanceofStub::Generate(MacroAssembler* masm) { |
__ Branch(&loop); |
__ bind(&is_instance); |
- ASSERT(Smi::FromInt(0) == 0); |
+ DCHECK(Smi::FromInt(0) == 0); |
if (!HasCallSiteInlineCheck()) { |
__ mov(v0, zero_reg); |
__ StoreRoot(v0, Heap::kInstanceofCacheAnswerRootIndex); |
@@ -1873,7 +1873,7 @@ void InstanceofStub::Generate(MacroAssembler* masm) { |
__ PatchRelocatedValue(inline_site, scratch, v0); |
if (!ReturnTrueFalseObject()) { |
- ASSERT_EQ(Smi::FromInt(0), 0); |
+ DCHECK_EQ(Smi::FromInt(0), 0); |
__ mov(v0, zero_reg); |
} |
} |
@@ -2093,7 +2093,7 @@ void ArgumentsAccessStub::GenerateNewSloppyFast(MacroAssembler* masm) { |
FixedArray::kHeaderSize + 2 * kPointerSize; |
// If there are no mapped parameters, we do not need the parameter_map. |
Label param_map_size; |
- ASSERT_EQ(0, Smi::FromInt(0)); |
+ DCHECK_EQ(0, Smi::FromInt(0)); |
__ Branch(USE_DELAY_SLOT, ¶m_map_size, eq, a1, Operand(zero_reg)); |
__ mov(t1, zero_reg); // In delay slot: param map size = 0 when a1 == 0. |
__ SmiScale(t1, a1, kPointerSizeLog2); |
@@ -2597,7 +2597,7 @@ void RegExpExecStub::Generate(MacroAssembler* masm) { |
__ li(a4, Operand( |
ExternalReference::address_of_static_offsets_vector(isolate()))); |
} else { // O32. |
- ASSERT(kMipsAbi == kO32); |
+ DCHECK(kMipsAbi == kO32); |
// Argument 9: Pass current isolate address. |
// CFunctionArgumentOperand handles MIPS stack argument slots. |
@@ -2848,9 +2848,9 @@ static void GenerateRecordCallTarget(MacroAssembler* masm) { |
// a3 : slot in feedback vector (Smi) |
Label initialize, done, miss, megamorphic, not_array_function; |
- ASSERT_EQ(*TypeFeedbackInfo::MegamorphicSentinel(masm->isolate()), |
+ DCHECK_EQ(*TypeFeedbackInfo::MegamorphicSentinel(masm->isolate()), |
masm->isolate()->heap()->megamorphic_symbol()); |
- ASSERT_EQ(*TypeFeedbackInfo::UninitializedSentinel(masm->isolate()), |
+ DCHECK_EQ(*TypeFeedbackInfo::UninitializedSentinel(masm->isolate()), |
masm->isolate()->heap()->uninitialized_symbol()); |
// Load the cache state into a4. |
@@ -3125,9 +3125,9 @@ void StringCharCodeAtGenerator::GenerateFast(MacroAssembler* masm) { |
Label got_char_code; |
Label sliced_string; |
- ASSERT(!a4.is(index_)); |
- ASSERT(!a4.is(result_)); |
- ASSERT(!a4.is(object_)); |
+ DCHECK(!a4.is(index_)); |
+ DCHECK(!a4.is(result_)); |
+ DCHECK(!a4.is(object_)); |
// If the receiver is a smi trigger the non-string case. |
__ JumpIfSmi(object_, receiver_not_string_); |
@@ -3326,7 +3326,7 @@ void StringCharCodeAtGenerator::GenerateSlow( |
if (index_flags_ == STRING_INDEX_IS_NUMBER) { |
__ CallRuntime(Runtime::kNumberToIntegerMapMinusZero, 1); |
} else { |
- ASSERT(index_flags_ == STRING_INDEX_IS_ARRAY_INDEX); |
+ DCHECK(index_flags_ == STRING_INDEX_IS_ARRAY_INDEX); |
// NumberToSmi discards numbers that are not exact integers. |
__ CallRuntime(Runtime::kNumberToSmi, 1); |
} |
@@ -3369,11 +3369,11 @@ void StringCharCodeAtGenerator::GenerateSlow( |
void StringCharFromCodeGenerator::GenerateFast(MacroAssembler* masm) { |
// Fast case of Heap::LookupSingleCharacterStringFromCode. |
- ASSERT(!a4.is(result_)); |
- ASSERT(!a4.is(code_)); |
+ DCHECK(!a4.is(result_)); |
+ DCHECK(!a4.is(code_)); |
STATIC_ASSERT(kSmiTag == 0); |
- ASSERT(IsPowerOf2(String::kMaxOneByteCharCode + 1)); |
+ DCHECK(IsPowerOf2(String::kMaxOneByteCharCode + 1)); |
__ And(a4, |
code_, |
Operand(kSmiTagMask | |
@@ -3758,7 +3758,7 @@ void StringCompareStub::GenerateFlatAsciiStringEquals(MacroAssembler* masm, |
__ bind(&check_zero_length); |
STATIC_ASSERT(kSmiTag == 0); |
__ Branch(&compare_chars, ne, length, Operand(zero_reg)); |
- ASSERT(is_int16((intptr_t)Smi::FromInt(EQUAL))); |
+ DCHECK(is_int16((intptr_t)Smi::FromInt(EQUAL))); |
__ Ret(USE_DELAY_SLOT); |
__ li(v0, Operand(Smi::FromInt(EQUAL))); |
@@ -3801,7 +3801,7 @@ void StringCompareStub::GenerateCompareFlatAsciiStrings(MacroAssembler* masm, |
// Compare lengths - strings up to min-length are equal. |
__ bind(&compare_lengths); |
- ASSERT(Smi::FromInt(EQUAL) == static_cast<Smi*>(0)); |
+ DCHECK(Smi::FromInt(EQUAL) == static_cast<Smi*>(0)); |
// Use length_delta as result if it's zero. |
__ mov(scratch2, length_delta); |
__ mov(scratch4, zero_reg); |
@@ -3917,7 +3917,7 @@ void BinaryOpICWithAllocationSiteStub::Generate(MacroAssembler* masm) { |
void ICCompareStub::GenerateSmis(MacroAssembler* masm) { |
- ASSERT(state_ == CompareIC::SMI); |
+ DCHECK(state_ == CompareIC::SMI); |
Label miss; |
__ Or(a2, a1, a0); |
__ JumpIfNotSmi(a2, &miss); |
@@ -3940,7 +3940,7 @@ void ICCompareStub::GenerateSmis(MacroAssembler* masm) { |
void ICCompareStub::GenerateNumbers(MacroAssembler* masm) { |
- ASSERT(state_ == CompareIC::NUMBER); |
+ DCHECK(state_ == CompareIC::NUMBER); |
Label generic_stub; |
Label unordered, maybe_undefined1, maybe_undefined2; |
@@ -3993,7 +3993,7 @@ void ICCompareStub::GenerateNumbers(MacroAssembler* masm) { |
__ BranchF(&fpu_lt, NULL, lt, f0, f2); |
// Otherwise it's greater, so just fall thru, and return. |
- ASSERT(is_int16(GREATER) && is_int16(EQUAL) && is_int16(LESS)); |
+ DCHECK(is_int16(GREATER) && is_int16(EQUAL) && is_int16(LESS)); |
__ Ret(USE_DELAY_SLOT); |
__ li(v0, Operand(GREATER)); |
@@ -4033,7 +4033,7 @@ void ICCompareStub::GenerateNumbers(MacroAssembler* masm) { |
void ICCompareStub::GenerateInternalizedStrings(MacroAssembler* masm) { |
- ASSERT(state_ == CompareIC::INTERNALIZED_STRING); |
+ DCHECK(state_ == CompareIC::INTERNALIZED_STRING); |
Label miss; |
// Registers containing left and right operands respectively. |
@@ -4057,13 +4057,13 @@ void ICCompareStub::GenerateInternalizedStrings(MacroAssembler* masm) { |
// Make sure a0 is non-zero. At this point input operands are |
// guaranteed to be non-zero. |
- ASSERT(right.is(a0)); |
+ DCHECK(right.is(a0)); |
STATIC_ASSERT(EQUAL == 0); |
STATIC_ASSERT(kSmiTag == 0); |
__ mov(v0, right); |
// Internalized strings are compared by identity. |
__ Ret(ne, left, Operand(right)); |
- ASSERT(is_int16(EQUAL)); |
+ DCHECK(is_int16(EQUAL)); |
__ Ret(USE_DELAY_SLOT); |
__ li(v0, Operand(Smi::FromInt(EQUAL))); |
@@ -4073,8 +4073,8 @@ void ICCompareStub::GenerateInternalizedStrings(MacroAssembler* masm) { |
void ICCompareStub::GenerateUniqueNames(MacroAssembler* masm) { |
- ASSERT(state_ == CompareIC::UNIQUE_NAME); |
- ASSERT(GetCondition() == eq); |
+ DCHECK(state_ == CompareIC::UNIQUE_NAME); |
+ DCHECK(GetCondition() == eq); |
Label miss; |
// Registers containing left and right operands respectively. |
@@ -4104,7 +4104,7 @@ void ICCompareStub::GenerateUniqueNames(MacroAssembler* masm) { |
__ Branch(&done, ne, left, Operand(right)); |
// Make sure a0 is non-zero. At this point input operands are |
// guaranteed to be non-zero. |
- ASSERT(right.is(a0)); |
+ DCHECK(right.is(a0)); |
STATIC_ASSERT(EQUAL == 0); |
STATIC_ASSERT(kSmiTag == 0); |
__ li(v0, Operand(Smi::FromInt(EQUAL))); |
@@ -4117,7 +4117,7 @@ void ICCompareStub::GenerateUniqueNames(MacroAssembler* masm) { |
void ICCompareStub::GenerateStrings(MacroAssembler* masm) { |
- ASSERT(state_ == CompareIC::STRING); |
+ DCHECK(state_ == CompareIC::STRING); |
Label miss; |
bool equality = Token::IsEqualityOp(op_); |
@@ -4160,7 +4160,7 @@ void ICCompareStub::GenerateStrings(MacroAssembler* masm) { |
// because we already know they are not identical. We know they are both |
// strings. |
if (equality) { |
- ASSERT(GetCondition() == eq); |
+ DCHECK(GetCondition() == eq); |
STATIC_ASSERT(kInternalizedTag == 0); |
__ Or(tmp3, tmp1, Operand(tmp2)); |
__ And(tmp5, tmp3, Operand(kIsNotInternalizedMask)); |
@@ -4168,7 +4168,7 @@ void ICCompareStub::GenerateStrings(MacroAssembler* masm) { |
__ Branch(&is_symbol, ne, tmp5, Operand(zero_reg)); |
// Make sure a0 is non-zero. At this point input operands are |
// guaranteed to be non-zero. |
- ASSERT(right.is(a0)); |
+ DCHECK(right.is(a0)); |
__ Ret(USE_DELAY_SLOT); |
__ mov(v0, a0); // In the delay slot. |
__ bind(&is_symbol); |
@@ -4203,7 +4203,7 @@ void ICCompareStub::GenerateStrings(MacroAssembler* masm) { |
void ICCompareStub::GenerateObjects(MacroAssembler* masm) { |
- ASSERT(state_ == CompareIC::OBJECT); |
+ DCHECK(state_ == CompareIC::OBJECT); |
Label miss; |
__ And(a2, a1, Operand(a0)); |
__ JumpIfSmi(a2, &miss); |
@@ -4213,7 +4213,7 @@ void ICCompareStub::GenerateObjects(MacroAssembler* masm) { |
__ GetObjectType(a1, a2, a2); |
__ Branch(&miss, ne, a2, Operand(JS_OBJECT_TYPE)); |
- ASSERT(GetCondition() == eq); |
+ DCHECK(GetCondition() == eq); |
__ Ret(USE_DELAY_SLOT); |
__ dsubu(v0, a0, a1); |
@@ -4302,7 +4302,7 @@ void NameDictionaryLookupStub::GenerateNegativeLookup(MacroAssembler* masm, |
Register properties, |
Handle<Name> name, |
Register scratch0) { |
- ASSERT(name->IsUniqueName()); |
+ DCHECK(name->IsUniqueName()); |
// If names of slots in range from 1 to kProbes - 1 for the hash value are |
// not equal to the name and kProbes-th slot is not used (its name is the |
// undefined value), it guarantees the hash table doesn't contain the |
@@ -4319,20 +4319,20 @@ void NameDictionaryLookupStub::GenerateNegativeLookup(MacroAssembler* masm, |
Operand(name->Hash() + NameDictionary::GetProbeOffset(i))); |
// Scale the index by multiplying by the entry size. |
- ASSERT(NameDictionary::kEntrySize == 3); |
+ DCHECK(NameDictionary::kEntrySize == 3); |
__ dsll(at, index, 1); |
__ Daddu(index, index, at); // index *= 3. |
Register entity_name = scratch0; |
// Having undefined at this place means the name is not contained. |
- ASSERT_EQ(kSmiTagSize, 1); |
+ DCHECK_EQ(kSmiTagSize, 1); |
Register tmp = properties; |
__ dsll(scratch0, index, kPointerSizeLog2); |
__ Daddu(tmp, properties, scratch0); |
__ ld(entity_name, FieldMemOperand(tmp, kElementsStartOffset)); |
- ASSERT(!tmp.is(entity_name)); |
+ DCHECK(!tmp.is(entity_name)); |
__ LoadRoot(tmp, Heap::kUndefinedValueRootIndex); |
__ Branch(done, eq, entity_name, Operand(tmp)); |
@@ -4385,10 +4385,10 @@ void NameDictionaryLookupStub::GeneratePositiveLookup(MacroAssembler* masm, |
Register name, |
Register scratch1, |
Register scratch2) { |
- ASSERT(!elements.is(scratch1)); |
- ASSERT(!elements.is(scratch2)); |
- ASSERT(!name.is(scratch1)); |
- ASSERT(!name.is(scratch2)); |
+ DCHECK(!elements.is(scratch1)); |
+ DCHECK(!elements.is(scratch2)); |
+ DCHECK(!name.is(scratch1)); |
+ DCHECK(!name.is(scratch2)); |
__ AssertName(name); |
@@ -4407,7 +4407,7 @@ void NameDictionaryLookupStub::GeneratePositiveLookup(MacroAssembler* masm, |
// Add the probe offset (i + i * i) left shifted to avoid right shifting |
// the hash in a separate instruction. The value hash + i + i * i is right |
// shifted in the following and instruction. |
- ASSERT(NameDictionary::GetProbeOffset(i) < |
+ DCHECK(NameDictionary::GetProbeOffset(i) < |
1 << (32 - Name::kHashFieldOffset)); |
__ Daddu(scratch2, scratch2, Operand( |
NameDictionary::GetProbeOffset(i) << Name::kHashShift)); |
@@ -4416,7 +4416,7 @@ void NameDictionaryLookupStub::GeneratePositiveLookup(MacroAssembler* masm, |
__ And(scratch2, scratch1, scratch2); |
// Scale the index by multiplying by the element size. |
- ASSERT(NameDictionary::kEntrySize == 3); |
+ DCHECK(NameDictionary::kEntrySize == 3); |
// scratch2 = scratch2 * 3. |
__ dsll(at, scratch2, 1); |
@@ -4436,7 +4436,7 @@ void NameDictionaryLookupStub::GeneratePositiveLookup(MacroAssembler* masm, |
__ MultiPush(spill_mask); |
if (name.is(a0)) { |
- ASSERT(!elements.is(a1)); |
+ DCHECK(!elements.is(a1)); |
__ Move(a1, name); |
__ Move(a0, elements); |
} else { |
@@ -4492,7 +4492,7 @@ void NameDictionaryLookupStub::Generate(MacroAssembler* masm) { |
// Add the probe offset (i + i * i) left shifted to avoid right shifting |
// the hash in a separate instruction. The value hash + i + i * i is right |
// shifted in the following and instruction. |
- ASSERT(NameDictionary::GetProbeOffset(i) < |
+ DCHECK(NameDictionary::GetProbeOffset(i) < |
1 << (32 - Name::kHashFieldOffset)); |
__ Daddu(index, hash, Operand( |
NameDictionary::GetProbeOffset(i) << Name::kHashShift)); |
@@ -4503,14 +4503,14 @@ void NameDictionaryLookupStub::Generate(MacroAssembler* masm) { |
__ And(index, mask, index); |
// Scale the index by multiplying by the entry size. |
- ASSERT(NameDictionary::kEntrySize == 3); |
+ DCHECK(NameDictionary::kEntrySize == 3); |
// index *= 3. |
__ mov(at, index); |
__ dsll(index, index, 1); |
__ Daddu(index, index, at); |
- ASSERT_EQ(kSmiTagSize, 1); |
+ DCHECK_EQ(kSmiTagSize, 1); |
__ dsll(index, index, kPointerSizeLog2); |
__ Daddu(index, index, dictionary); |
__ ld(entry_key, FieldMemOperand(index, kElementsStartOffset)); |
@@ -4647,8 +4647,8 @@ void RecordWriteStub::InformIncrementalMarker(MacroAssembler* masm) { |
__ PrepareCallCFunction(argument_count, regs_.scratch0()); |
Register address = |
a0.is(regs_.address()) ? regs_.scratch0() : regs_.address(); |
- ASSERT(!address.is(regs_.object())); |
- ASSERT(!address.is(a0)); |
+ DCHECK(!address.is(regs_.object())); |
+ DCHECK(!address.is(a0)); |
__ Move(address, regs_.address()); |
__ Move(a0, regs_.object()); |
__ Move(a1, address); |
@@ -4869,7 +4869,7 @@ void ProfileEntryHookStub::Generate(MacroAssembler* masm) { |
int frame_alignment = masm->ActivationFrameAlignment(); |
if (frame_alignment > kPointerSize) { |
__ mov(s5, sp); |
- ASSERT(IsPowerOf2(frame_alignment)); |
+ DCHECK(IsPowerOf2(frame_alignment)); |
__ And(sp, sp, Operand(-frame_alignment)); |
} |
@@ -4937,12 +4937,12 @@ static void CreateArrayDispatchOneArgument(MacroAssembler* masm, |
// sp[0] - last argument |
Label normal_sequence; |
if (mode == DONT_OVERRIDE) { |
- ASSERT(FAST_SMI_ELEMENTS == 0); |
- ASSERT(FAST_HOLEY_SMI_ELEMENTS == 1); |
- ASSERT(FAST_ELEMENTS == 2); |
- ASSERT(FAST_HOLEY_ELEMENTS == 3); |
- ASSERT(FAST_DOUBLE_ELEMENTS == 4); |
- ASSERT(FAST_HOLEY_DOUBLE_ELEMENTS == 5); |
+ DCHECK(FAST_SMI_ELEMENTS == 0); |
+ DCHECK(FAST_HOLEY_SMI_ELEMENTS == 1); |
+ DCHECK(FAST_ELEMENTS == 2); |
+ DCHECK(FAST_HOLEY_ELEMENTS == 3); |
+ DCHECK(FAST_DOUBLE_ELEMENTS == 4); |
+ DCHECK(FAST_HOLEY_DOUBLE_ELEMENTS == 5); |
// is the low bit set? If so, we are holey and that is good. |
__ And(at, a3, Operand(1)); |
@@ -5249,7 +5249,7 @@ void CallApiFunctionStub::Generate(MacroAssembler* masm) { |
FrameScope frame_scope(masm, StackFrame::MANUAL); |
__ EnterExitFrame(false, kApiStackSpace); |
- ASSERT(!api_function_address.is(a0) && !scratch.is(a0)); |
+ DCHECK(!api_function_address.is(a0) && !scratch.is(a0)); |
// a0 = FunctionCallbackInfo& |
// Arguments is after the return address. |
__ Daddu(a0, sp, Operand(1 * kPointerSize)); |