| Index: src/lookup-cache.cc
|
| diff --git a/src/lookup-cache.cc b/src/lookup-cache.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..18729d630d5b0068d0273c315c2d81c58c4fb995
|
| --- /dev/null
|
| +++ b/src/lookup-cache.cc
|
| @@ -0,0 +1,84 @@
|
| +// Copyright 2016 the V8 project authors. All rights reserved.
|
| +// Use of this source code is governed by a BSD-style license that can be
|
| +// found in the LICENSE file.
|
| +
|
| +#include "src/lookup-cache.h"
|
| +
|
| +#include "src/objects-inl.h"
|
| +
|
| +namespace v8 {
|
| +namespace internal {
|
| +
|
| +void DescriptorLookupCache::Clear() {
|
| + for (int index = 0; index < kLength; index++) keys_[index].source = NULL;
|
| +}
|
| +
|
| +int KeyedLookupCache::Hash(Handle<Map> map, Handle<Name> name) {
|
| + DisallowHeapAllocation no_gc;
|
| + // Uses only lower 32 bits if pointers are larger.
|
| + uintptr_t addr_hash =
|
| + static_cast<uint32_t>(reinterpret_cast<uintptr_t>(*map)) >> kMapHashShift;
|
| + return static_cast<uint32_t>((addr_hash ^ name->Hash()) & kCapacityMask);
|
| +}
|
| +
|
| +int KeyedLookupCache::Lookup(Handle<Map> map, Handle<Name> name) {
|
| + DisallowHeapAllocation no_gc;
|
| + int index = (Hash(map, name) & kHashMask);
|
| + for (int i = 0; i < kEntriesPerBucket; i++) {
|
| + Key& key = keys_[index + i];
|
| + if ((key.map == *map) && key.name->Equals(*name)) {
|
| + return field_offsets_[index + i];
|
| + }
|
| + }
|
| + return kNotFound;
|
| +}
|
| +
|
| +void KeyedLookupCache::Update(Handle<Map> map, Handle<Name> name,
|
| + int field_offset) {
|
| + DisallowHeapAllocation no_gc;
|
| + if (!name->IsUniqueName()) {
|
| + if (!StringTable::InternalizeStringIfExists(name->GetIsolate(),
|
| + Handle<String>::cast(name))
|
| + .ToHandle(&name)) {
|
| + return;
|
| + }
|
| + }
|
| + // This cache is cleared only between mark compact passes, so we expect the
|
| + // cache to only contain old space names.
|
| + DCHECK(!map->GetIsolate()->heap()->InNewSpace(*name));
|
| +
|
| + int index = (Hash(map, name) & kHashMask);
|
| + // After a GC there will be free slots, so we use them in order (this may
|
| + // help to get the most frequently used one in position 0).
|
| + for (int i = 0; i < kEntriesPerBucket; i++) {
|
| + Key& key = keys_[index];
|
| + Object* free_entry_indicator = NULL;
|
| + if (key.map == free_entry_indicator) {
|
| + key.map = *map;
|
| + key.name = *name;
|
| + field_offsets_[index + i] = field_offset;
|
| + return;
|
| + }
|
| + }
|
| + // No free entry found in this bucket, so we move them all down one and
|
| + // put the new entry at position zero.
|
| + for (int i = kEntriesPerBucket - 1; i > 0; i--) {
|
| + Key& key = keys_[index + i];
|
| + Key& key2 = keys_[index + i - 1];
|
| + key = key2;
|
| + field_offsets_[index + i] = field_offsets_[index + i - 1];
|
| + }
|
| +
|
| + // Write the new first entry.
|
| + Key& key = keys_[index];
|
| + key.map = *map;
|
| + key.name = *name;
|
| + field_offsets_[index] = field_offset;
|
| +}
|
| +
|
| +void KeyedLookupCache::Clear() {
|
| + for (int index = 0; index < kLength; index++) keys_[index].map = NULL;
|
| +}
|
| +
|
| +} // namespace internal
|
| +} // namespace v8
|
|
|