| Index: src/lookup-cache.cc
 | 
| diff --git a/src/lookup-cache.cc b/src/lookup-cache.cc
 | 
| index 18729d630d5b0068d0273c315c2d81c58c4fb995..b740fdbf110f141a1894ea5ae7f50d4e8d1a30f5 100644
 | 
| --- a/src/lookup-cache.cc
 | 
| +++ b/src/lookup-cache.cc
 | 
| @@ -13,72 +13,5 @@ void DescriptorLookupCache::Clear() {
 | 
|    for (int index = 0; index < kLength; index++) keys_[index].source = NULL;
 | 
|  }
 | 
|  
 | 
| -int KeyedLookupCache::Hash(Handle<Map> map, Handle<Name> name) {
 | 
| -  DisallowHeapAllocation no_gc;
 | 
| -  // Uses only lower 32 bits if pointers are larger.
 | 
| -  uintptr_t addr_hash =
 | 
| -      static_cast<uint32_t>(reinterpret_cast<uintptr_t>(*map)) >> kMapHashShift;
 | 
| -  return static_cast<uint32_t>((addr_hash ^ name->Hash()) & kCapacityMask);
 | 
| -}
 | 
| -
 | 
| -int KeyedLookupCache::Lookup(Handle<Map> map, Handle<Name> name) {
 | 
| -  DisallowHeapAllocation no_gc;
 | 
| -  int index = (Hash(map, name) & kHashMask);
 | 
| -  for (int i = 0; i < kEntriesPerBucket; i++) {
 | 
| -    Key& key = keys_[index + i];
 | 
| -    if ((key.map == *map) && key.name->Equals(*name)) {
 | 
| -      return field_offsets_[index + i];
 | 
| -    }
 | 
| -  }
 | 
| -  return kNotFound;
 | 
| -}
 | 
| -
 | 
| -void KeyedLookupCache::Update(Handle<Map> map, Handle<Name> name,
 | 
| -                              int field_offset) {
 | 
| -  DisallowHeapAllocation no_gc;
 | 
| -  if (!name->IsUniqueName()) {
 | 
| -    if (!StringTable::InternalizeStringIfExists(name->GetIsolate(),
 | 
| -                                                Handle<String>::cast(name))
 | 
| -             .ToHandle(&name)) {
 | 
| -      return;
 | 
| -    }
 | 
| -  }
 | 
| -  // This cache is cleared only between mark compact passes, so we expect the
 | 
| -  // cache to only contain old space names.
 | 
| -  DCHECK(!map->GetIsolate()->heap()->InNewSpace(*name));
 | 
| -
 | 
| -  int index = (Hash(map, name) & kHashMask);
 | 
| -  // After a GC there will be free slots, so we use them in order (this may
 | 
| -  // help to get the most frequently used one in position 0).
 | 
| -  for (int i = 0; i < kEntriesPerBucket; i++) {
 | 
| -    Key& key = keys_[index];
 | 
| -    Object* free_entry_indicator = NULL;
 | 
| -    if (key.map == free_entry_indicator) {
 | 
| -      key.map = *map;
 | 
| -      key.name = *name;
 | 
| -      field_offsets_[index + i] = field_offset;
 | 
| -      return;
 | 
| -    }
 | 
| -  }
 | 
| -  // No free entry found in this bucket, so we move them all down one and
 | 
| -  // put the new entry at position zero.
 | 
| -  for (int i = kEntriesPerBucket - 1; i > 0; i--) {
 | 
| -    Key& key = keys_[index + i];
 | 
| -    Key& key2 = keys_[index + i - 1];
 | 
| -    key = key2;
 | 
| -    field_offsets_[index + i] = field_offsets_[index + i - 1];
 | 
| -  }
 | 
| -
 | 
| -  // Write the new first entry.
 | 
| -  Key& key = keys_[index];
 | 
| -  key.map = *map;
 | 
| -  key.name = *name;
 | 
| -  field_offsets_[index] = field_offset;
 | 
| -}
 | 
| -
 | 
| -void KeyedLookupCache::Clear() {
 | 
| -  for (int index = 0; index < kLength; index++) keys_[index].map = NULL;
 | 
| -}
 | 
| -
 | 
|  }  // namespace internal
 | 
|  }  // namespace v8
 | 
| 
 |