Index: src/ic/arm/stub-cache-arm.cc |
diff --git a/src/ic/arm/stub-cache-arm.cc b/src/ic/arm/stub-cache-arm.cc |
index 3dba9a30c4f6188a95b26775ed2d2261cc1e4cc8..f87c324cbd4bff96e42ccdd92c5f1ed7fffb1831 100644 |
--- a/src/ic/arm/stub-cache-arm.cc |
+++ b/src/ic/arm/stub-cache-arm.cc |
@@ -15,8 +15,7 @@ namespace internal { |
#define __ ACCESS_MASM(masm) |
static void ProbeTable(StubCache* stub_cache, MacroAssembler* masm, |
- Code::Flags flags, StubCache::Table table, |
- Register receiver, Register name, |
+ StubCache::Table table, Register receiver, Register name, |
// Number of the cache entry, not scaled. |
Register offset, Register scratch, Register scratch2, |
Register offset_scratch) { |
@@ -69,13 +68,15 @@ static void ProbeTable(StubCache* stub_cache, MacroAssembler* masm, |
__ ldr(flags_reg, FieldMemOperand(code, Code::kFlagsOffset)); |
// It's a nice optimization if this constant is encodable in the bic insn. |
+#ifdef DEBUG |
+ Code::Flags flags = Code::RemoveHolderFromFlags( |
+ Code::ComputeHandlerFlags(stub_cache->ic_kind())); |
uint32_t mask = Code::kFlagsNotUsedInLookup; |
DCHECK(__ ImmediateFitsAddrMode1Instruction(mask)); |
__ bic(flags_reg, flags_reg, Operand(mask)); |
__ cmp(flags_reg, Operand(flags)); |
- __ b(ne, &miss); |
+ __ Check(eq, kUnexpectedValue); |
-#ifdef DEBUG |
if (FLAG_test_secondary_stub_cache && table == StubCache::kPrimary) { |
__ jmp(&miss); |
} else if (FLAG_test_primary_stub_cache && table == StubCache::kSecondary) { |
@@ -93,9 +94,6 @@ static void ProbeTable(StubCache* stub_cache, MacroAssembler* masm, |
void StubCache::GenerateProbe(MacroAssembler* masm, Register receiver, |
Register name, Register scratch, Register extra, |
Register extra2, Register extra3) { |
- Code::Flags flags = |
- Code::RemoveHolderFromFlags(Code::ComputeHandlerFlags(ic_kind_)); |
- |
Label miss; |
// Make sure that code is valid. The multiplying code relies on the |
@@ -144,25 +142,21 @@ void StubCache::GenerateProbe(MacroAssembler* masm, Register receiver, |
// We shift out the last two bits because they are not part of the hash and |
// they are always 01 for maps. |
__ mov(scratch, Operand(scratch, LSR, kCacheIndexShift)); |
- // Mask down the eor argument to the minimum to keep the immediate |
- // ARM-encodable. |
- __ eor(scratch, scratch, Operand((flags >> kCacheIndexShift) & mask)); |
// Prefer and_ to ubfx here because ubfx takes 2 cycles. |
__ and_(scratch, scratch, Operand(mask)); |
// Probe the primary table. |
- ProbeTable(this, masm, flags, kPrimary, receiver, name, scratch, extra, |
- extra2, extra3); |
+ ProbeTable(this, masm, kPrimary, receiver, name, scratch, extra, extra2, |
+ extra3); |
// Primary miss: Compute hash for secondary probe. |
__ sub(scratch, scratch, Operand(name, LSR, kCacheIndexShift)); |
uint32_t mask2 = kSecondaryTableSize - 1; |
- __ add(scratch, scratch, Operand((flags >> kCacheIndexShift) & mask2)); |
__ and_(scratch, scratch, Operand(mask2)); |
// Probe the secondary table. |
- ProbeTable(this, masm, flags, kSecondary, receiver, name, scratch, extra, |
- extra2, extra3); |
+ ProbeTable(this, masm, kSecondary, receiver, name, scratch, extra, extra2, |
+ extra3); |
// Cache miss: Fall-through and let caller handle the miss by |
// entering the runtime system. |