| Index: third_party/WebKit/Source/wtf/HashTable.h
|
| diff --git a/third_party/WebKit/Source/wtf/HashTable.h b/third_party/WebKit/Source/wtf/HashTable.h
|
| index 85b68a728df8094a6e2eebe56044ed159f838323..1f8584bd62e79aff2a0414901177d485f2a4b192 100644
|
| --- a/third_party/WebKit/Source/wtf/HashTable.h
|
| +++ b/third_party/WebKit/Source/wtf/HashTable.h
|
| @@ -36,35 +36,39 @@
|
|
|
| #if DUMP_HASHTABLE_STATS
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| -#define UPDATE_PROBE_COUNTS() \
|
| - ++probeCount; \
|
| - HashTableStats::recordCollisionAtCount(probeCount); \
|
| - ++perTableProbeCount; \
|
| - m_stats->recordCollisionAtCount(perTableProbeCount)
|
| -#define UPDATE_ACCESS_COUNTS() \
|
| - atomicIncrement(&HashTableStats::numAccesses); \
|
| - int probeCount = 0; \
|
| - ++m_stats->numAccesses; \
|
| - int perTableProbeCount = 0
|
| +#define UPDATE_PROBE_COUNTS() \
|
| + ++probeCount; \
|
| + HashTableStats::recordCollisionAtCount(probeCount); \
|
| + ++perTableProbeCount; \
|
| + m_stats->recordCollisionAtCount(perTableProbeCount)
|
| +#define UPDATE_ACCESS_COUNTS() \
|
| + atomicIncrement(&HashTableStats::numAccesses); \
|
| + int probeCount = 0; \
|
| + ++m_stats->numAccesses; \
|
| + int perTableProbeCount = 0
|
| #else
|
| -#define UPDATE_PROBE_COUNTS() \
|
| - ++probeCount; \
|
| - HashTableStats::recordCollisionAtCount(probeCount)
|
| -#define UPDATE_ACCESS_COUNTS() \
|
| - atomicIncrement(&HashTableStats::numAccesses); \
|
| - int probeCount = 0
|
| +#define UPDATE_PROBE_COUNTS() \
|
| + ++probeCount; \
|
| + HashTableStats::recordCollisionAtCount(probeCount)
|
| +#define UPDATE_ACCESS_COUNTS() \
|
| + atomicIncrement(&HashTableStats::numAccesses); \
|
| + int probeCount = 0
|
| #endif
|
| #else
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| -#define UPDATE_PROBE_COUNTS() \
|
| - ++perTableProbeCount; \
|
| - m_stats->recordCollisionAtCount(perTableProbeCount)
|
| -#define UPDATE_ACCESS_COUNTS() \
|
| - ++m_stats->numAccesses; \
|
| - int perTableProbeCount = 0
|
| +#define UPDATE_PROBE_COUNTS() \
|
| + ++perTableProbeCount; \
|
| + m_stats->recordCollisionAtCount(perTableProbeCount)
|
| +#define UPDATE_ACCESS_COUNTS() \
|
| + ++m_stats->numAccesses; \
|
| + int perTableProbeCount = 0
|
| #else
|
| -#define UPDATE_PROBE_COUNTS() do { } while (0)
|
| -#define UPDATE_ACCESS_COUNTS() do { } while (0)
|
| +#define UPDATE_PROBE_COUNTS() \
|
| + do { \
|
| + } while (0)
|
| +#define UPDATE_ACCESS_COUNTS() \
|
| + do { \
|
| + } while (0)
|
| #endif
|
| #endif
|
|
|
| @@ -73,1445 +77,2043 @@ namespace WTF {
|
| #if DUMP_HASHTABLE_STATS
|
|
|
| struct HashTableStats {
|
| - STATIC_ONLY(HashTableStats);
|
| - // The following variables are all atomically incremented when modified.
|
| - static int numAccesses;
|
| - static int numRehashes;
|
| - static int numRemoves;
|
| - static int numReinserts;
|
| -
|
| - // The following variables are only modified in the recordCollisionAtCount
|
| - // method within a mutex.
|
| - static int maxCollisions;
|
| - static int numCollisions;
|
| - static int collisionGraph[4096];
|
| -
|
| - static void recordCollisionAtCount(int count);
|
| - static void dumpStats();
|
| + STATIC_ONLY(HashTableStats);
|
| + // The following variables are all atomically incremented when modified.
|
| + static int numAccesses;
|
| + static int numRehashes;
|
| + static int numRemoves;
|
| + static int numReinserts;
|
| +
|
| + // The following variables are only modified in the recordCollisionAtCount
|
| + // method within a mutex.
|
| + static int maxCollisions;
|
| + static int numCollisions;
|
| + static int collisionGraph[4096];
|
| +
|
| + static void recordCollisionAtCount(int count);
|
| + static void dumpStats();
|
| };
|
|
|
| #endif
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| class HashTable;
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| class HashTableIterator;
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| class HashTableConstIterator;
|
| -template <typename Value, typename HashFunctions, typename HashTraits, typename Allocator>
|
| +template <typename Value,
|
| + typename HashFunctions,
|
| + typename HashTraits,
|
| + typename Allocator>
|
| class LinkedHashSet;
|
| -template <WeakHandlingFlag x, typename T, typename U, typename V, typename W, typename X, typename Y, typename Z>
|
| +template <WeakHandlingFlag x,
|
| + typename T,
|
| + typename U,
|
| + typename V,
|
| + typename W,
|
| + typename X,
|
| + typename Y,
|
| + typename Z>
|
| struct WeakProcessingHashTableHelper;
|
|
|
| typedef enum { HashItemKnownGood } HashItemKnownGoodTag;
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| class HashTableConstIterator final {
|
| - DISALLOW_NEW();
|
| -private:
|
| - typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> HashTableType;
|
| - typedef HashTableIterator<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> iterator;
|
| - typedef HashTableConstIterator<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> const_iterator;
|
| - typedef Value ValueType;
|
| - typedef typename Traits::IteratorConstGetType GetType;
|
| - typedef const ValueType* PointerType;
|
| -
|
| - friend class HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>;
|
| - friend class HashTableIterator<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>;
|
| -
|
| - void skipEmptyBuckets()
|
| - {
|
| - while (m_position != m_endPosition && HashTableType::isEmptyOrDeletedBucket(*m_position))
|
| - ++m_position;
|
| - }
|
| -
|
| - HashTableConstIterator(PointerType position, PointerType endPosition, const HashTableType* container)
|
| - : m_position(position)
|
| - , m_endPosition(endPosition)
|
| + DISALLOW_NEW();
|
| +
|
| + private:
|
| + typedef HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + HashTableType;
|
| + typedef HashTableIterator<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + iterator;
|
| + typedef HashTableConstIterator<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + const_iterator;
|
| + typedef Value ValueType;
|
| + typedef typename Traits::IteratorConstGetType GetType;
|
| + typedef const ValueType* PointerType;
|
| +
|
| + friend class HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>;
|
| + friend class HashTableIterator<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>;
|
| +
|
| + void skipEmptyBuckets() {
|
| + while (m_position != m_endPosition &&
|
| + HashTableType::isEmptyOrDeletedBucket(*m_position))
|
| + ++m_position;
|
| + }
|
| +
|
| + HashTableConstIterator(PointerType position,
|
| + PointerType endPosition,
|
| + const HashTableType* container)
|
| + : m_position(position),
|
| + m_endPosition(endPosition)
|
| #if ENABLE(ASSERT)
|
| - , m_container(container)
|
| - , m_containerModifications(container->modifications())
|
| + ,
|
| + m_container(container),
|
| + m_containerModifications(container->modifications())
|
| #endif
|
| - {
|
| - skipEmptyBuckets();
|
| - }
|
| -
|
| - HashTableConstIterator(PointerType position, PointerType endPosition, const HashTableType* container, HashItemKnownGoodTag)
|
| - : m_position(position)
|
| - , m_endPosition(endPosition)
|
| + {
|
| + skipEmptyBuckets();
|
| + }
|
| +
|
| + HashTableConstIterator(PointerType position,
|
| + PointerType endPosition,
|
| + const HashTableType* container,
|
| + HashItemKnownGoodTag)
|
| + : m_position(position),
|
| + m_endPosition(endPosition)
|
| #if ENABLE(ASSERT)
|
| - , m_container(container)
|
| - , m_containerModifications(container->modifications())
|
| + ,
|
| + m_container(container),
|
| + m_containerModifications(container->modifications())
|
| #endif
|
| - {
|
| - ASSERT(m_containerModifications == m_container->modifications());
|
| - }
|
| -
|
| - void checkModifications() const
|
| - {
|
| - // HashTable and collections that build on it do not support
|
| - // modifications while there is an iterator in use. The exception is
|
| - // ListHashSet, which has its own iterators that tolerate modification
|
| - // of the underlying set.
|
| - ASSERT(m_containerModifications == m_container->modifications());
|
| - ASSERT(!m_container->accessForbidden());
|
| - }
|
| -
|
| -public:
|
| - HashTableConstIterator() {}
|
| -
|
| - GetType get() const
|
| - {
|
| - checkModifications();
|
| - return m_position;
|
| - }
|
| - typename Traits::IteratorConstReferenceType operator*() const { return Traits::getToReferenceConstConversion(get()); }
|
| - GetType operator->() const { return get(); }
|
| -
|
| - const_iterator& operator++()
|
| - {
|
| - ASSERT(m_position != m_endPosition);
|
| - checkModifications();
|
| - ++m_position;
|
| - skipEmptyBuckets();
|
| - return *this;
|
| - }
|
| -
|
| - // postfix ++ intentionally omitted
|
| -
|
| - // Comparison.
|
| - bool operator==(const const_iterator& other) const
|
| - {
|
| - return m_position == other.m_position;
|
| - }
|
| - bool operator!=(const const_iterator& other) const
|
| - {
|
| - return m_position != other.m_position;
|
| - }
|
| - bool operator==(const iterator& other) const
|
| - {
|
| - return *this == static_cast<const_iterator>(other);
|
| - }
|
| - bool operator!=(const iterator& other) const
|
| - {
|
| - return *this != static_cast<const_iterator>(other);
|
| - }
|
| -
|
| -private:
|
| - PointerType m_position;
|
| - PointerType m_endPosition;
|
| + {
|
| + ASSERT(m_containerModifications == m_container->modifications());
|
| + }
|
| +
|
| + void checkModifications() const {
|
| + // HashTable and collections that build on it do not support
|
| + // modifications while there is an iterator in use. The exception is
|
| + // ListHashSet, which has its own iterators that tolerate modification
|
| + // of the underlying set.
|
| + ASSERT(m_containerModifications == m_container->modifications());
|
| + ASSERT(!m_container->accessForbidden());
|
| + }
|
| +
|
| + public:
|
| + HashTableConstIterator() {}
|
| +
|
| + GetType get() const {
|
| + checkModifications();
|
| + return m_position;
|
| + }
|
| + typename Traits::IteratorConstReferenceType operator*() const {
|
| + return Traits::getToReferenceConstConversion(get());
|
| + }
|
| + GetType operator->() const { return get(); }
|
| +
|
| + const_iterator& operator++() {
|
| + ASSERT(m_position != m_endPosition);
|
| + checkModifications();
|
| + ++m_position;
|
| + skipEmptyBuckets();
|
| + return *this;
|
| + }
|
| +
|
| + // postfix ++ intentionally omitted
|
| +
|
| + // Comparison.
|
| + bool operator==(const const_iterator& other) const {
|
| + return m_position == other.m_position;
|
| + }
|
| + bool operator!=(const const_iterator& other) const {
|
| + return m_position != other.m_position;
|
| + }
|
| + bool operator==(const iterator& other) const {
|
| + return *this == static_cast<const_iterator>(other);
|
| + }
|
| + bool operator!=(const iterator& other) const {
|
| + return *this != static_cast<const_iterator>(other);
|
| + }
|
| +
|
| + private:
|
| + PointerType m_position;
|
| + PointerType m_endPosition;
|
| #if ENABLE(ASSERT)
|
| - const HashTableType* m_container;
|
| - int64_t m_containerModifications;
|
| + const HashTableType* m_container;
|
| + int64_t m_containerModifications;
|
| #endif
|
| };
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| class HashTableIterator final {
|
| - DISALLOW_NEW();
|
| -private:
|
| - typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> HashTableType;
|
| - typedef HashTableIterator<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> iterator;
|
| - typedef HashTableConstIterator<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> const_iterator;
|
| - typedef Value ValueType;
|
| - typedef typename Traits::IteratorGetType GetType;
|
| - typedef ValueType* PointerType;
|
| -
|
| - friend class HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>;
|
| -
|
| - HashTableIterator(PointerType pos, PointerType end, const HashTableType* container) : m_iterator(pos, end, container) {}
|
| - HashTableIterator(PointerType pos, PointerType end, const HashTableType* container, HashItemKnownGoodTag tag) : m_iterator(pos, end, container, tag) {}
|
| -
|
| -public:
|
| - HashTableIterator() {}
|
| -
|
| - // default copy, assignment and destructor are OK
|
| -
|
| - GetType get() const { return const_cast<GetType>(m_iterator.get()); }
|
| - typename Traits::IteratorReferenceType operator*() const { return Traits::getToReferenceConversion(get()); }
|
| - GetType operator->() const { return get(); }
|
| -
|
| - iterator& operator++() { ++m_iterator; return *this; }
|
| -
|
| - // postfix ++ intentionally omitted
|
| -
|
| - // Comparison.
|
| - bool operator==(const iterator& other) const { return m_iterator == other.m_iterator; }
|
| - bool operator!=(const iterator& other) const { return m_iterator != other.m_iterator; }
|
| - bool operator==(const const_iterator& other) const { return m_iterator == other; }
|
| - bool operator!=(const const_iterator& other) const { return m_iterator != other; }
|
| -
|
| - operator const_iterator() const { return m_iterator; }
|
| -
|
| -private:
|
| - const_iterator m_iterator;
|
| + DISALLOW_NEW();
|
| +
|
| + private:
|
| + typedef HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + HashTableType;
|
| + typedef HashTableIterator<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + iterator;
|
| + typedef HashTableConstIterator<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + const_iterator;
|
| + typedef Value ValueType;
|
| + typedef typename Traits::IteratorGetType GetType;
|
| + typedef ValueType* PointerType;
|
| +
|
| + friend class HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>;
|
| +
|
| + HashTableIterator(PointerType pos,
|
| + PointerType end,
|
| + const HashTableType* container)
|
| + : m_iterator(pos, end, container) {}
|
| + HashTableIterator(PointerType pos,
|
| + PointerType end,
|
| + const HashTableType* container,
|
| + HashItemKnownGoodTag tag)
|
| + : m_iterator(pos, end, container, tag) {}
|
| +
|
| + public:
|
| + HashTableIterator() {}
|
| +
|
| + // default copy, assignment and destructor are OK
|
| +
|
| + GetType get() const { return const_cast<GetType>(m_iterator.get()); }
|
| + typename Traits::IteratorReferenceType operator*() const {
|
| + return Traits::getToReferenceConversion(get());
|
| + }
|
| + GetType operator->() const { return get(); }
|
| +
|
| + iterator& operator++() {
|
| + ++m_iterator;
|
| + return *this;
|
| + }
|
| +
|
| + // postfix ++ intentionally omitted
|
| +
|
| + // Comparison.
|
| + bool operator==(const iterator& other) const {
|
| + return m_iterator == other.m_iterator;
|
| + }
|
| + bool operator!=(const iterator& other) const {
|
| + return m_iterator != other.m_iterator;
|
| + }
|
| + bool operator==(const const_iterator& other) const {
|
| + return m_iterator == other;
|
| + }
|
| + bool operator!=(const const_iterator& other) const {
|
| + return m_iterator != other;
|
| + }
|
| +
|
| + operator const_iterator() const { return m_iterator; }
|
| +
|
| + private:
|
| + const_iterator m_iterator;
|
| };
|
|
|
| using std::swap;
|
|
|
| // Work around MSVC's standard library, whose swap for pairs does not swap by component.
|
| -template <typename T> inline void hashTableSwap(T& a, T& b)
|
| -{
|
| - swap(a, b);
|
| +template <typename T>
|
| +inline void hashTableSwap(T& a, T& b) {
|
| + swap(a, b);
|
| }
|
|
|
| -template <typename T, typename U> inline void hashTableSwap(KeyValuePair<T, U>& a, KeyValuePair<T, U>& b)
|
| -{
|
| - swap(a.key, b.key);
|
| - swap(a.value, b.value);
|
| +template <typename T, typename U>
|
| +inline void hashTableSwap(KeyValuePair<T, U>& a, KeyValuePair<T, U>& b) {
|
| + swap(a.key, b.key);
|
| + swap(a.value, b.value);
|
| }
|
|
|
| -template <typename T, typename Allocator, bool useSwap = !IsTriviallyDestructible<T>::value>
|
| +template <typename T,
|
| + typename Allocator,
|
| + bool useSwap = !IsTriviallyDestructible<T>::value>
|
| struct Mover;
|
| -template <typename T, typename Allocator> struct Mover<T, Allocator, true> {
|
| - STATIC_ONLY(Mover);
|
| - static void move(T& from, T& to)
|
| - {
|
| - // The key and value cannot be swapped atomically, and it would be wrong
|
| - // to have a GC when only one was swapped and the other still contained
|
| - // garbage (eg. from a previous use of the same slot). Therefore we
|
| - // forbid a GC until both the key and the value are swapped.
|
| - Allocator::enterGCForbiddenScope();
|
| - hashTableSwap(from, to);
|
| - Allocator::leaveGCForbiddenScope();
|
| - }
|
| +template <typename T, typename Allocator>
|
| +struct Mover<T, Allocator, true> {
|
| + STATIC_ONLY(Mover);
|
| + static void move(T& from, T& to) {
|
| + // The key and value cannot be swapped atomically, and it would be wrong
|
| + // to have a GC when only one was swapped and the other still contained
|
| + // garbage (eg. from a previous use of the same slot). Therefore we
|
| + // forbid a GC until both the key and the value are swapped.
|
| + Allocator::enterGCForbiddenScope();
|
| + hashTableSwap(from, to);
|
| + Allocator::leaveGCForbiddenScope();
|
| + }
|
| };
|
|
|
| -template <typename T, typename Allocator> struct Mover<T, Allocator, false> {
|
| - STATIC_ONLY(Mover);
|
| - static void move(T& from, T& to) { to = from; }
|
| +template <typename T, typename Allocator>
|
| +struct Mover<T, Allocator, false> {
|
| + STATIC_ONLY(Mover);
|
| + static void move(T& from, T& to) { to = from; }
|
| };
|
|
|
| -template <typename HashFunctions> class IdentityHashTranslator {
|
| - STATIC_ONLY(IdentityHashTranslator);
|
| -public:
|
| - template <typename T> static unsigned hash(const T& key) { return HashFunctions::hash(key); }
|
| - template <typename T, typename U> static bool equal(const T& a, const U& b) { return HashFunctions::equal(a, b); }
|
| - template <typename T, typename U, typename V> static void translate(T& location, const U&, const V& value) { location = value; }
|
| +template <typename HashFunctions>
|
| +class IdentityHashTranslator {
|
| + STATIC_ONLY(IdentityHashTranslator);
|
| +
|
| + public:
|
| + template <typename T>
|
| + static unsigned hash(const T& key) {
|
| + return HashFunctions::hash(key);
|
| + }
|
| + template <typename T, typename U>
|
| + static bool equal(const T& a, const U& b) {
|
| + return HashFunctions::equal(a, b);
|
| + }
|
| + template <typename T, typename U, typename V>
|
| + static void translate(T& location, const U&, const V& value) {
|
| + location = value;
|
| + }
|
| };
|
|
|
| -template <typename HashTableType, typename ValueType> struct HashTableAddResult final {
|
| - STACK_ALLOCATED();
|
| - HashTableAddResult(const HashTableType* container, ValueType* storedValue, bool isNewEntry)
|
| - : storedValue(storedValue)
|
| - , isNewEntry(isNewEntry)
|
| +template <typename HashTableType, typename ValueType>
|
| +struct HashTableAddResult final {
|
| + STACK_ALLOCATED();
|
| + HashTableAddResult(const HashTableType* container,
|
| + ValueType* storedValue,
|
| + bool isNewEntry)
|
| + : storedValue(storedValue),
|
| + isNewEntry(isNewEntry)
|
| #if ENABLE(SECURITY_ASSERT)
|
| - , m_container(container)
|
| - , m_containerModifications(container->modifications())
|
| + ,
|
| + m_container(container),
|
| + m_containerModifications(container->modifications())
|
| #endif
|
| - {
|
| - ASSERT_UNUSED(container, container);
|
| - }
|
| + {
|
| + ASSERT_UNUSED(container, container);
|
| + }
|
|
|
| - ValueType* storedValue;
|
| - bool isNewEntry;
|
| + ValueType* storedValue;
|
| + bool isNewEntry;
|
|
|
| #if ENABLE(SECURITY_ASSERT)
|
| - ~HashTableAddResult()
|
| - {
|
| - // If rehash happened before accessing storedValue, it's
|
| - // use-after-free. Any modification may cause a rehash, so we check for
|
| - // modifications here.
|
| -
|
| - // Rehash after accessing storedValue is harmless but will assert if the
|
| - // AddResult destructor takes place after a modification. You may need
|
| - // to limit the scope of the AddResult.
|
| - ASSERT_WITH_SECURITY_IMPLICATION(m_containerModifications == m_container->modifications());
|
| - }
|
| -
|
| -private:
|
| - const HashTableType* m_container;
|
| - const int64_t m_containerModifications;
|
| + ~HashTableAddResult() {
|
| + // If rehash happened before accessing storedValue, it's
|
| + // use-after-free. Any modification may cause a rehash, so we check for
|
| + // modifications here.
|
| +
|
| + // Rehash after accessing storedValue is harmless but will assert if the
|
| + // AddResult destructor takes place after a modification. You may need
|
| + // to limit the scope of the AddResult.
|
| + ASSERT_WITH_SECURITY_IMPLICATION(m_containerModifications ==
|
| + m_container->modifications());
|
| + }
|
| +
|
| + private:
|
| + const HashTableType* m_container;
|
| + const int64_t m_containerModifications;
|
| #endif
|
| };
|
|
|
| template <typename Value, typename Extractor, typename KeyTraits>
|
| struct HashTableHelper {
|
| - STATIC_ONLY(HashTableHelper);
|
| - static bool isEmptyBucket(const Value& value) { return isHashTraitsEmptyValue<KeyTraits>(Extractor::extract(value)); }
|
| - static bool isDeletedBucket(const Value& value) { return KeyTraits::isDeletedValue(Extractor::extract(value)); }
|
| - static bool isEmptyOrDeletedBucket(const Value& value) { return isEmptyBucket(value) || isDeletedBucket(value); }
|
| + STATIC_ONLY(HashTableHelper);
|
| + static bool isEmptyBucket(const Value& value) {
|
| + return isHashTraitsEmptyValue<KeyTraits>(Extractor::extract(value));
|
| + }
|
| + static bool isDeletedBucket(const Value& value) {
|
| + return KeyTraits::isDeletedValue(Extractor::extract(value));
|
| + }
|
| + static bool isEmptyOrDeletedBucket(const Value& value) {
|
| + return isEmptyBucket(value) || isDeletedBucket(value);
|
| + }
|
| };
|
|
|
| -template <typename HashTranslator, typename KeyTraits, bool safeToCompareToEmptyOrDeleted>
|
| +template <typename HashTranslator,
|
| + typename KeyTraits,
|
| + bool safeToCompareToEmptyOrDeleted>
|
| struct HashTableKeyChecker {
|
| - STATIC_ONLY(HashTableKeyChecker);
|
| - // There's no simple generic way to make this check if
|
| - // safeToCompareToEmptyOrDeleted is false, so the check always passes.
|
| - template <typename T>
|
| - static bool checkKey(const T&) { return true; }
|
| + STATIC_ONLY(HashTableKeyChecker);
|
| + // There's no simple generic way to make this check if
|
| + // safeToCompareToEmptyOrDeleted is false, so the check always passes.
|
| + template <typename T>
|
| + static bool checkKey(const T&) {
|
| + return true;
|
| + }
|
| };
|
|
|
| template <typename HashTranslator, typename KeyTraits>
|
| struct HashTableKeyChecker<HashTranslator, KeyTraits, true> {
|
| - STATIC_ONLY(HashTableKeyChecker);
|
| - template <typename T>
|
| - static bool checkKey(const T& key)
|
| - {
|
| - // FIXME : Check also equality to the deleted value.
|
| - return !HashTranslator::equal(KeyTraits::emptyValue(), key);
|
| - }
|
| + STATIC_ONLY(HashTableKeyChecker);
|
| + template <typename T>
|
| + static bool checkKey(const T& key) {
|
| + // FIXME : Check also equality to the deleted value.
|
| + return !HashTranslator::equal(KeyTraits::emptyValue(), key);
|
| + }
|
| };
|
|
|
| // Note: empty or deleted key values are not allowed, using them may lead to
|
| // undefined behavior. For pointer keys this means that null pointers are not
|
| // allowed unless you supply custom key traits.
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -class HashTable final : public ConditionalDestructor<HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>, Allocator::isGarbageCollected> {
|
| - DISALLOW_NEW();
|
| -public:
|
| - typedef HashTableIterator<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> iterator;
|
| - typedef HashTableConstIterator<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> const_iterator;
|
| - typedef Traits ValueTraits;
|
| - typedef Key KeyType;
|
| - typedef typename KeyTraits::PeekInType KeyPeekInType;
|
| - typedef typename KeyTraits::PassInType KeyPassInType;
|
| - typedef Value ValueType;
|
| - typedef Extractor ExtractorType;
|
| - typedef KeyTraits KeyTraitsType;
|
| - typedef typename Traits::PassInType ValuePassInType;
|
| - typedef IdentityHashTranslator<HashFunctions> IdentityTranslatorType;
|
| - typedef HashTableAddResult<HashTable, ValueType> AddResult;
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +class HashTable final
|
| + : public ConditionalDestructor<HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>,
|
| + Allocator::isGarbageCollected> {
|
| + DISALLOW_NEW();
|
| +
|
| + public:
|
| + typedef HashTableIterator<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + iterator;
|
| + typedef HashTableConstIterator<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>
|
| + const_iterator;
|
| + typedef Traits ValueTraits;
|
| + typedef Key KeyType;
|
| + typedef typename KeyTraits::PeekInType KeyPeekInType;
|
| + typedef typename KeyTraits::PassInType KeyPassInType;
|
| + typedef Value ValueType;
|
| + typedef Extractor ExtractorType;
|
| + typedef KeyTraits KeyTraitsType;
|
| + typedef typename Traits::PassInType ValuePassInType;
|
| + typedef IdentityHashTranslator<HashFunctions> IdentityTranslatorType;
|
| + typedef HashTableAddResult<HashTable, ValueType> AddResult;
|
|
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - struct Stats {
|
| - DISALLOW_NEW(Stats);
|
| - Stats()
|
| - : numAccesses(0)
|
| - , numRehashes(0)
|
| - , numRemoves(0)
|
| - , numReinserts(0)
|
| - , maxCollisions(0)
|
| - , numCollisions(0)
|
| - , collisionGraph()
|
| - {
|
| - }
|
| -
|
| - int numAccesses;
|
| - int numRehashes;
|
| - int numRemoves;
|
| - int numReinserts;
|
| -
|
| - int maxCollisions;
|
| - int numCollisions;
|
| - int collisionGraph[4096];
|
| -
|
| - void recordCollisionAtCount(int count)
|
| - {
|
| - if (count > maxCollisions)
|
| - maxCollisions = count;
|
| - numCollisions++;
|
| - collisionGraph[count]++;
|
| - }
|
| + struct Stats {
|
| + DISALLOW_NEW(Stats);
|
| + Stats()
|
| + : numAccesses(0),
|
| + numRehashes(0),
|
| + numRemoves(0),
|
| + numReinserts(0),
|
| + maxCollisions(0),
|
| + numCollisions(0),
|
| + collisionGraph() {}
|
| +
|
| + int numAccesses;
|
| + int numRehashes;
|
| + int numRemoves;
|
| + int numReinserts;
|
| +
|
| + int maxCollisions;
|
| + int numCollisions;
|
| + int collisionGraph[4096];
|
| +
|
| + void recordCollisionAtCount(int count) {
|
| + if (count > maxCollisions)
|
| + maxCollisions = count;
|
| + numCollisions++;
|
| + collisionGraph[count]++;
|
| + }
|
|
|
| - void dumpStats()
|
| - {
|
| - dataLogF("\nWTF::HashTable::Stats dump\n\n");
|
| - dataLogF("%d accesses\n", numAccesses);
|
| - dataLogF("%d total collisions, average %.2f probes per access\n", numCollisions, 1.0 * (numAccesses + numCollisions) / numAccesses);
|
| - dataLogF("longest collision chain: %d\n", maxCollisions);
|
| - for (int i = 1; i <= maxCollisions; i++) {
|
| - dataLogF(" %d lookups with exactly %d collisions (%.2f%% , %.2f%% with this many or more)\n", collisionGraph[i], i, 100.0 * (collisionGraph[i] - collisionGraph[i+1]) / numAccesses, 100.0 * collisionGraph[i] / numAccesses);
|
| - }
|
| - dataLogF("%d rehashes\n", numRehashes);
|
| - dataLogF("%d reinserts\n", numReinserts);
|
| - }
|
| - };
|
| + void dumpStats() {
|
| + dataLogF("\nWTF::HashTable::Stats dump\n\n");
|
| + dataLogF("%d accesses\n", numAccesses);
|
| + dataLogF("%d total collisions, average %.2f probes per access\n",
|
| + numCollisions,
|
| + 1.0 * (numAccesses + numCollisions) / numAccesses);
|
| + dataLogF("longest collision chain: %d\n", maxCollisions);
|
| + for (int i = 1; i <= maxCollisions; i++) {
|
| + dataLogF(
|
| + " %d lookups with exactly %d collisions (%.2f%% , %.2f%% with "
|
| + "this many or more)\n",
|
| + collisionGraph[i], i,
|
| + 100.0 * (collisionGraph[i] - collisionGraph[i + 1]) / numAccesses,
|
| + 100.0 * collisionGraph[i] / numAccesses);
|
| + }
|
| + dataLogF("%d rehashes\n", numRehashes);
|
| + dataLogF("%d reinserts\n", numReinserts);
|
| + }
|
| + };
|
| #endif
|
|
|
| - HashTable();
|
| - void finalize()
|
| - {
|
| - ASSERT(!Allocator::isGarbageCollected);
|
| - if (LIKELY(!m_table))
|
| - return;
|
| - ASSERT(!m_accessForbidden);
|
| + HashTable();
|
| + void finalize() {
|
| + ASSERT(!Allocator::isGarbageCollected);
|
| + if (LIKELY(!m_table))
|
| + return;
|
| + ASSERT(!m_accessForbidden);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = true;
|
| + m_accessForbidden = true;
|
| #endif
|
| - deleteAllBucketsAndDeallocate(m_table, m_tableSize);
|
| + deleteAllBucketsAndDeallocate(m_table, m_tableSize);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = false;
|
| + m_accessForbidden = false;
|
| #endif
|
| - m_table = nullptr;
|
| - }
|
| -
|
| - HashTable(const HashTable&);
|
| - void swap(HashTable&);
|
| - HashTable& operator=(const HashTable&);
|
| -
|
| - // When the hash table is empty, just return the same iterator for end as
|
| - // for begin. This is more efficient because we don't have to skip all the
|
| - // empty and deleted buckets, and iterating an empty table is a common case
|
| - // that's worth optimizing.
|
| - iterator begin() { return isEmpty() ? end() : makeIterator(m_table); }
|
| - iterator end() { return makeKnownGoodIterator(m_table + m_tableSize); }
|
| - const_iterator begin() const { return isEmpty() ? end() : makeConstIterator(m_table); }
|
| - const_iterator end() const { return makeKnownGoodConstIterator(m_table + m_tableSize); }
|
| -
|
| - unsigned size() const
|
| - {
|
| - ASSERT(!m_accessForbidden);
|
| - return m_keyCount;
|
| - }
|
| - unsigned capacity() const
|
| - {
|
| - ASSERT(!m_accessForbidden);
|
| - return m_tableSize;
|
| - }
|
| - bool isEmpty() const
|
| - {
|
| - ASSERT(!m_accessForbidden);
|
| - return !m_keyCount;
|
| - }
|
| -
|
| - void reserveCapacityForSize(unsigned size);
|
| -
|
| - AddResult add(ValuePassInType value)
|
| - {
|
| - return add<IdentityTranslatorType>(Extractor::extract(value), value);
|
| - }
|
| -
|
| - // A special version of add() that finds the object by hashing and comparing
|
| - // with some other type, to avoid the cost of type conversion if the object
|
| - // is already in the table.
|
| - template <typename HashTranslator, typename T, typename Extra> AddResult add(const T& key, const Extra&);
|
| - template <typename HashTranslator, typename T, typename Extra> AddResult addPassingHashCode(const T& key, const Extra&);
|
| -
|
| - iterator find(KeyPeekInType key) { return find<IdentityTranslatorType>(key); }
|
| - const_iterator find(KeyPeekInType key) const { return find<IdentityTranslatorType>(key); }
|
| - bool contains(KeyPeekInType key) const { return contains<IdentityTranslatorType>(key); }
|
| -
|
| - template <typename HashTranslator, typename T> iterator find(const T&);
|
| - template <typename HashTranslator, typename T> const_iterator find(const T&) const;
|
| - template <typename HashTranslator, typename T> bool contains(const T&) const;
|
| -
|
| - void remove(KeyPeekInType);
|
| - void remove(iterator);
|
| - void remove(const_iterator);
|
| - void clear();
|
| -
|
| - static bool isEmptyBucket(const ValueType& value) { return isHashTraitsEmptyValue<KeyTraits>(Extractor::extract(value)); }
|
| - static bool isDeletedBucket(const ValueType& value) { return KeyTraits::isDeletedValue(Extractor::extract(value)); }
|
| - static bool isEmptyOrDeletedBucket(const ValueType& value) { return HashTableHelper<ValueType, Extractor, KeyTraits>:: isEmptyOrDeletedBucket(value); }
|
| -
|
| - ValueType* lookup(KeyPeekInType key) { return lookup<IdentityTranslatorType, KeyPeekInType>(key); }
|
| - template <typename HashTranslator, typename T> ValueType* lookup(T);
|
| - template <typename HashTranslator, typename T> const ValueType* lookup(T) const;
|
| -
|
| - template <typename VisitorDispatcher> void trace(VisitorDispatcher);
|
| + m_table = nullptr;
|
| + }
|
| +
|
| + HashTable(const HashTable&);
|
| + void swap(HashTable&);
|
| + HashTable& operator=(const HashTable&);
|
| +
|
| + // When the hash table is empty, just return the same iterator for end as
|
| + // for begin. This is more efficient because we don't have to skip all the
|
| + // empty and deleted buckets, and iterating an empty table is a common case
|
| + // that's worth optimizing.
|
| + iterator begin() { return isEmpty() ? end() : makeIterator(m_table); }
|
| + iterator end() { return makeKnownGoodIterator(m_table + m_tableSize); }
|
| + const_iterator begin() const {
|
| + return isEmpty() ? end() : makeConstIterator(m_table);
|
| + }
|
| + const_iterator end() const {
|
| + return makeKnownGoodConstIterator(m_table + m_tableSize);
|
| + }
|
| +
|
| + unsigned size() const {
|
| + ASSERT(!m_accessForbidden);
|
| + return m_keyCount;
|
| + }
|
| + unsigned capacity() const {
|
| + ASSERT(!m_accessForbidden);
|
| + return m_tableSize;
|
| + }
|
| + bool isEmpty() const {
|
| + ASSERT(!m_accessForbidden);
|
| + return !m_keyCount;
|
| + }
|
| +
|
| + void reserveCapacityForSize(unsigned size);
|
| +
|
| + AddResult add(ValuePassInType value) {
|
| + return add<IdentityTranslatorType>(Extractor::extract(value), value);
|
| + }
|
| +
|
| + // A special version of add() that finds the object by hashing and comparing
|
| + // with some other type, to avoid the cost of type conversion if the object
|
| + // is already in the table.
|
| + template <typename HashTranslator, typename T, typename Extra>
|
| + AddResult add(const T& key, const Extra&);
|
| + template <typename HashTranslator, typename T, typename Extra>
|
| + AddResult addPassingHashCode(const T& key, const Extra&);
|
| +
|
| + iterator find(KeyPeekInType key) { return find<IdentityTranslatorType>(key); }
|
| + const_iterator find(KeyPeekInType key) const {
|
| + return find<IdentityTranslatorType>(key);
|
| + }
|
| + bool contains(KeyPeekInType key) const {
|
| + return contains<IdentityTranslatorType>(key);
|
| + }
|
| +
|
| + template <typename HashTranslator, typename T>
|
| + iterator find(const T&);
|
| + template <typename HashTranslator, typename T>
|
| + const_iterator find(const T&) const;
|
| + template <typename HashTranslator, typename T>
|
| + bool contains(const T&) const;
|
| +
|
| + void remove(KeyPeekInType);
|
| + void remove(iterator);
|
| + void remove(const_iterator);
|
| + void clear();
|
| +
|
| + static bool isEmptyBucket(const ValueType& value) {
|
| + return isHashTraitsEmptyValue<KeyTraits>(Extractor::extract(value));
|
| + }
|
| + static bool isDeletedBucket(const ValueType& value) {
|
| + return KeyTraits::isDeletedValue(Extractor::extract(value));
|
| + }
|
| + static bool isEmptyOrDeletedBucket(const ValueType& value) {
|
| + return HashTableHelper<ValueType, Extractor,
|
| + KeyTraits>::isEmptyOrDeletedBucket(value);
|
| + }
|
| +
|
| + ValueType* lookup(KeyPeekInType key) {
|
| + return lookup<IdentityTranslatorType, KeyPeekInType>(key);
|
| + }
|
| + template <typename HashTranslator, typename T>
|
| + ValueType* lookup(T);
|
| + template <typename HashTranslator, typename T>
|
| + const ValueType* lookup(T) const;
|
| +
|
| + template <typename VisitorDispatcher>
|
| + void trace(VisitorDispatcher);
|
|
|
| #if ENABLE(ASSERT)
|
| - bool accessForbidden() const { return m_accessForbidden; }
|
| - int64_t modifications() const { return m_modifications; }
|
| - void registerModification() { m_modifications++; }
|
| - // HashTable and collections that build on it do not support modifications
|
| - // while there is an iterator in use. The exception is ListHashSet, which
|
| - // has its own iterators that tolerate modification of the underlying set.
|
| - void checkModifications(int64_t mods) const { ASSERT(mods == m_modifications); }
|
| + bool accessForbidden() const { return m_accessForbidden; }
|
| + int64_t modifications() const { return m_modifications; }
|
| + void registerModification() { m_modifications++; }
|
| + // HashTable and collections that build on it do not support modifications
|
| + // while there is an iterator in use. The exception is ListHashSet, which
|
| + // has its own iterators that tolerate modification of the underlying set.
|
| + void checkModifications(int64_t mods) const {
|
| + ASSERT(mods == m_modifications);
|
| + }
|
| #else
|
| - int64_t modifications() const { return 0; }
|
| - void registerModification() {}
|
| - void checkModifications(int64_t mods) const {}
|
| + int64_t modifications() const { return 0; }
|
| + void registerModification() {}
|
| + void checkModifications(int64_t mods) const {}
|
| #endif
|
|
|
| -private:
|
| - static ValueType* allocateTable(unsigned size);
|
| - static void deleteAllBucketsAndDeallocate(ValueType* table, unsigned size);
|
| -
|
| - typedef std::pair<ValueType*, bool> LookupType;
|
| - typedef std::pair<LookupType, unsigned> FullLookupType;
|
| -
|
| - LookupType lookupForWriting(const Key& key) { return lookupForWriting<IdentityTranslatorType>(key); }
|
| - template <typename HashTranslator, typename T> FullLookupType fullLookupForWriting(const T&);
|
| - template <typename HashTranslator, typename T> LookupType lookupForWriting(const T&);
|
| -
|
| - void remove(ValueType*);
|
| -
|
| - bool shouldExpand() const { return (m_keyCount + m_deletedCount) * m_maxLoad >= m_tableSize; }
|
| - bool mustRehashInPlace() const { return m_keyCount * m_minLoad < m_tableSize * 2; }
|
| - bool shouldShrink() const
|
| - {
|
| - // isAllocationAllowed check should be at the last because it's
|
| - // expensive.
|
| - return m_keyCount * m_minLoad < m_tableSize
|
| - && m_tableSize > KeyTraits::minimumTableSize
|
| - && Allocator::isAllocationAllowed();
|
| - }
|
| - ValueType* expand(ValueType* entry = 0);
|
| - void shrink() { rehash(m_tableSize / 2, 0); }
|
| -
|
| - ValueType* expandBuffer(unsigned newTableSize, ValueType* entry, bool&);
|
| - ValueType* rehashTo(ValueType* newTable, unsigned newTableSize, ValueType* entry);
|
| - ValueType* rehash(unsigned newTableSize, ValueType* entry);
|
| - ValueType* reinsert(ValueType&);
|
| -
|
| - static void initializeBucket(ValueType& bucket);
|
| - static void deleteBucket(ValueType& bucket)
|
| - {
|
| - bucket.~ValueType();
|
| - Traits::constructDeletedValue(bucket, Allocator::isGarbageCollected);
|
| - }
|
| -
|
| - FullLookupType makeLookupResult(ValueType* position, bool found, unsigned hash)
|
| - { return FullLookupType(LookupType(position, found), hash); }
|
| -
|
| - iterator makeIterator(ValueType* pos) { return iterator(pos, m_table + m_tableSize, this); }
|
| - const_iterator makeConstIterator(ValueType* pos) const { return const_iterator(pos, m_table + m_tableSize, this); }
|
| - iterator makeKnownGoodIterator(ValueType* pos) { return iterator(pos, m_table + m_tableSize, this, HashItemKnownGood); }
|
| - const_iterator makeKnownGoodConstIterator(ValueType* pos) const { return const_iterator(pos, m_table + m_tableSize, this, HashItemKnownGood); }
|
| -
|
| - static const unsigned m_maxLoad = 2;
|
| - static const unsigned m_minLoad = 6;
|
| -
|
| - unsigned tableSizeMask() const
|
| - {
|
| - size_t mask = m_tableSize - 1;
|
| - ASSERT((mask & m_tableSize) == 0);
|
| - return mask;
|
| - }
|
| -
|
| - void setEnqueued() { m_queueFlag = true; }
|
| - void clearEnqueued() { m_queueFlag = false; }
|
| - bool enqueued() { return m_queueFlag; }
|
| -
|
| - ValueType* m_table;
|
| - unsigned m_tableSize;
|
| - unsigned m_keyCount;
|
| + private:
|
| + static ValueType* allocateTable(unsigned size);
|
| + static void deleteAllBucketsAndDeallocate(ValueType* table, unsigned size);
|
| +
|
| + typedef std::pair<ValueType*, bool> LookupType;
|
| + typedef std::pair<LookupType, unsigned> FullLookupType;
|
| +
|
| + LookupType lookupForWriting(const Key& key) {
|
| + return lookupForWriting<IdentityTranslatorType>(key);
|
| + }
|
| + template <typename HashTranslator, typename T>
|
| + FullLookupType fullLookupForWriting(const T&);
|
| + template <typename HashTranslator, typename T>
|
| + LookupType lookupForWriting(const T&);
|
| +
|
| + void remove(ValueType*);
|
| +
|
| + bool shouldExpand() const {
|
| + return (m_keyCount + m_deletedCount) * m_maxLoad >= m_tableSize;
|
| + }
|
| + bool mustRehashInPlace() const {
|
| + return m_keyCount * m_minLoad < m_tableSize * 2;
|
| + }
|
| + bool shouldShrink() const {
|
| + // isAllocationAllowed check should be at the last because it's
|
| + // expensive.
|
| + return m_keyCount * m_minLoad < m_tableSize &&
|
| + m_tableSize > KeyTraits::minimumTableSize &&
|
| + Allocator::isAllocationAllowed();
|
| + }
|
| + ValueType* expand(ValueType* entry = 0);
|
| + void shrink() { rehash(m_tableSize / 2, 0); }
|
| +
|
| + ValueType* expandBuffer(unsigned newTableSize, ValueType* entry, bool&);
|
| + ValueType* rehashTo(ValueType* newTable,
|
| + unsigned newTableSize,
|
| + ValueType* entry);
|
| + ValueType* rehash(unsigned newTableSize, ValueType* entry);
|
| + ValueType* reinsert(ValueType&);
|
| +
|
| + static void initializeBucket(ValueType& bucket);
|
| + static void deleteBucket(ValueType& bucket) {
|
| + bucket.~ValueType();
|
| + Traits::constructDeletedValue(bucket, Allocator::isGarbageCollected);
|
| + }
|
| +
|
| + FullLookupType makeLookupResult(ValueType* position,
|
| + bool found,
|
| + unsigned hash) {
|
| + return FullLookupType(LookupType(position, found), hash);
|
| + }
|
| +
|
| + iterator makeIterator(ValueType* pos) {
|
| + return iterator(pos, m_table + m_tableSize, this);
|
| + }
|
| + const_iterator makeConstIterator(ValueType* pos) const {
|
| + return const_iterator(pos, m_table + m_tableSize, this);
|
| + }
|
| + iterator makeKnownGoodIterator(ValueType* pos) {
|
| + return iterator(pos, m_table + m_tableSize, this, HashItemKnownGood);
|
| + }
|
| + const_iterator makeKnownGoodConstIterator(ValueType* pos) const {
|
| + return const_iterator(pos, m_table + m_tableSize, this, HashItemKnownGood);
|
| + }
|
| +
|
| + static const unsigned m_maxLoad = 2;
|
| + static const unsigned m_minLoad = 6;
|
| +
|
| + unsigned tableSizeMask() const {
|
| + size_t mask = m_tableSize - 1;
|
| + ASSERT((mask & m_tableSize) == 0);
|
| + return mask;
|
| + }
|
| +
|
| + void setEnqueued() { m_queueFlag = true; }
|
| + void clearEnqueued() { m_queueFlag = false; }
|
| + bool enqueued() { return m_queueFlag; }
|
| +
|
| + ValueType* m_table;
|
| + unsigned m_tableSize;
|
| + unsigned m_keyCount;
|
| #if ENABLE(ASSERT)
|
| - unsigned m_deletedCount:30;
|
| - unsigned m_queueFlag:1;
|
| - unsigned m_accessForbidden:1;
|
| - unsigned m_modifications;
|
| + unsigned m_deletedCount : 30;
|
| + unsigned m_queueFlag : 1;
|
| + unsigned m_accessForbidden : 1;
|
| + unsigned m_modifications;
|
| #else
|
| - unsigned m_deletedCount:31;
|
| - unsigned m_queueFlag:1;
|
| + unsigned m_deletedCount : 31;
|
| + unsigned m_queueFlag : 1;
|
| #endif
|
|
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| -public:
|
| - mutable OwnPtr<Stats> m_stats;
|
| + public:
|
| + mutable OwnPtr<Stats> m_stats;
|
| #endif
|
|
|
| - template <WeakHandlingFlag x, typename T, typename U, typename V, typename W, typename X, typename Y, typename Z> friend struct WeakProcessingHashTableHelper;
|
| - template <typename T, typename U, typename V, typename W> friend class LinkedHashSet;
|
| + template <WeakHandlingFlag x,
|
| + typename T,
|
| + typename U,
|
| + typename V,
|
| + typename W,
|
| + typename X,
|
| + typename Y,
|
| + typename Z>
|
| + friend struct WeakProcessingHashTableHelper;
|
| + template <typename T, typename U, typename V, typename W>
|
| + friend class LinkedHashSet;
|
| };
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -inline HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::HashTable()
|
| - : m_table(nullptr)
|
| - , m_tableSize(0)
|
| - , m_keyCount(0)
|
| - , m_deletedCount(0)
|
| - , m_queueFlag(false)
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +inline HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::HashTable()
|
| + : m_table(nullptr),
|
| + m_tableSize(0),
|
| + m_keyCount(0),
|
| + m_deletedCount(0),
|
| + m_queueFlag(false)
|
| #if ENABLE(ASSERT)
|
| - , m_accessForbidden(false)
|
| - , m_modifications(0)
|
| + ,
|
| + m_accessForbidden(false),
|
| + m_modifications(0)
|
| #endif
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - , m_stats(adoptPtr(new Stats))
|
| + ,
|
| + m_stats(adoptPtr(new Stats))
|
| #endif
|
| {
|
| - static_assert(Allocator::isGarbageCollected || (!IsPointerToGarbageCollectedType<Key>::value && !IsPointerToGarbageCollectedType<Value>::value), "Cannot put raw pointers to garbage-collected classes into an off-heap collection.");
|
| + static_assert(Allocator::isGarbageCollected ||
|
| + (!IsPointerToGarbageCollectedType<Key>::value &&
|
| + !IsPointerToGarbageCollectedType<Value>::value),
|
| + "Cannot put raw pointers to garbage-collected classes into an "
|
| + "off-heap collection.");
|
| }
|
|
|
| -inline unsigned doubleHash(unsigned key)
|
| -{
|
| - key = ~key + (key >> 23);
|
| - key ^= (key << 12);
|
| - key ^= (key >> 7);
|
| - key ^= (key << 2);
|
| - key ^= (key >> 20);
|
| - return key;
|
| +inline unsigned doubleHash(unsigned key) {
|
| + key = ~key + (key >> 23);
|
| + key ^= (key << 12);
|
| + key ^= (key >> 7);
|
| + key ^= (key << 2);
|
| + key ^= (key >> 20);
|
| + return key;
|
| }
|
|
|
| -inline unsigned calculateCapacity(unsigned size)
|
| -{
|
| - for (unsigned mask = size; mask; mask >>= 1)
|
| - size |= mask; // 00110101010 -> 00111111111
|
| - return (size + 1) * 2; // 00111111111 -> 10000000000
|
| +inline unsigned calculateCapacity(unsigned size) {
|
| + for (unsigned mask = size; mask; mask >>= 1)
|
| + size |= mask; // 00110101010 -> 00111111111
|
| + return (size + 1) * 2; // 00111111111 -> 10000000000
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::reserveCapacityForSize(unsigned newSize)
|
| -{
|
| - unsigned newCapacity = calculateCapacity(newSize);
|
| - if (newCapacity < KeyTraits::minimumTableSize)
|
| - newCapacity = KeyTraits::minimumTableSize;
|
| -
|
| - if (newCapacity > capacity()) {
|
| - RELEASE_ASSERT(!static_cast<int>(newCapacity >> 31)); // HashTable capacity should not overflow 32bit int.
|
| - rehash(newCapacity, 0);
|
| - }
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +void HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::reserveCapacityForSize(unsigned newSize) {
|
| + unsigned newCapacity = calculateCapacity(newSize);
|
| + if (newCapacity < KeyTraits::minimumTableSize)
|
| + newCapacity = KeyTraits::minimumTableSize;
|
| +
|
| + if (newCapacity > capacity()) {
|
| + RELEASE_ASSERT(!static_cast<int>(
|
| + newCapacity >>
|
| + 31)); // HashTable capacity should not overflow 32bit int.
|
| + rehash(newCapacity, 0);
|
| + }
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T>
|
| -inline Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::lookup(T key)
|
| -{
|
| - return const_cast<Value*>(const_cast<const HashTable*>(this)->lookup<HashTranslator, T>(key));
|
| +inline Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + lookup(T key) {
|
| + return const_cast<Value*>(
|
| + const_cast<const HashTable*>(this)->lookup<HashTranslator, T>(key));
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T>
|
| -inline const Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::lookup(T key) const
|
| -{
|
| - ASSERT(!m_accessForbidden);
|
| - ASSERT((HashTableKeyChecker<HashTranslator, KeyTraits, HashFunctions::safeToCompareToEmptyOrDeleted>::checkKey(key)));
|
| - const ValueType* table = m_table;
|
| - if (!table)
|
| +inline const Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + lookup(T key) const {
|
| + ASSERT(!m_accessForbidden);
|
| + ASSERT((HashTableKeyChecker<
|
| + HashTranslator, KeyTraits,
|
| + HashFunctions::safeToCompareToEmptyOrDeleted>::checkKey(key)));
|
| + const ValueType* table = m_table;
|
| + if (!table)
|
| + return nullptr;
|
| +
|
| + size_t k = 0;
|
| + size_t sizeMask = tableSizeMask();
|
| + unsigned h = HashTranslator::hash(key);
|
| + size_t i = h & sizeMask;
|
| +
|
| + UPDATE_ACCESS_COUNTS();
|
| +
|
| + while (1) {
|
| + const ValueType* entry = table + i;
|
| +
|
| + if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| + if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return entry;
|
| +
|
| + if (isEmptyBucket(*entry))
|
| + return nullptr;
|
| + } else {
|
| + if (isEmptyBucket(*entry))
|
| return nullptr;
|
|
|
| - size_t k = 0;
|
| - size_t sizeMask = tableSizeMask();
|
| - unsigned h = HashTranslator::hash(key);
|
| - size_t i = h & sizeMask;
|
| -
|
| - UPDATE_ACCESS_COUNTS();
|
| -
|
| - while (1) {
|
| - const ValueType* entry = table + i;
|
| -
|
| - if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| - if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return entry;
|
| -
|
| - if (isEmptyBucket(*entry))
|
| - return nullptr;
|
| - } else {
|
| - if (isEmptyBucket(*entry))
|
| - return nullptr;
|
| -
|
| - if (!isDeletedBucket(*entry) && HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return entry;
|
| - }
|
| - UPDATE_PROBE_COUNTS();
|
| - if (!k)
|
| - k = 1 | doubleHash(h);
|
| - i = (i + k) & sizeMask;
|
| + if (!isDeletedBucket(*entry) &&
|
| + HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return entry;
|
| }
|
| + UPDATE_PROBE_COUNTS();
|
| + if (!k)
|
| + k = 1 | doubleHash(h);
|
| + i = (i + k) & sizeMask;
|
| + }
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T>
|
| -inline typename HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::LookupType HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::lookupForWriting(const T& key)
|
| -{
|
| - ASSERT(!m_accessForbidden);
|
| - ASSERT(m_table);
|
| - registerModification();
|
| -
|
| - ValueType* table = m_table;
|
| - size_t k = 0;
|
| - size_t sizeMask = tableSizeMask();
|
| - unsigned h = HashTranslator::hash(key);
|
| - size_t i = h & sizeMask;
|
| -
|
| - UPDATE_ACCESS_COUNTS();
|
| -
|
| - ValueType* deletedEntry = nullptr;
|
| -
|
| - while (1) {
|
| - ValueType* entry = table + i;
|
| -
|
| - if (isEmptyBucket(*entry))
|
| - return LookupType(deletedEntry ? deletedEntry : entry, false);
|
| -
|
| - if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| - if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return LookupType(entry, true);
|
| -
|
| - if (isDeletedBucket(*entry))
|
| - deletedEntry = entry;
|
| - } else {
|
| - if (isDeletedBucket(*entry))
|
| - deletedEntry = entry;
|
| - else if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return LookupType(entry, true);
|
| - }
|
| - UPDATE_PROBE_COUNTS();
|
| - if (!k)
|
| - k = 1 | doubleHash(h);
|
| - i = (i + k) & sizeMask;
|
| +inline typename HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::LookupType
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + lookupForWriting(const T& key) {
|
| + ASSERT(!m_accessForbidden);
|
| + ASSERT(m_table);
|
| + registerModification();
|
| +
|
| + ValueType* table = m_table;
|
| + size_t k = 0;
|
| + size_t sizeMask = tableSizeMask();
|
| + unsigned h = HashTranslator::hash(key);
|
| + size_t i = h & sizeMask;
|
| +
|
| + UPDATE_ACCESS_COUNTS();
|
| +
|
| + ValueType* deletedEntry = nullptr;
|
| +
|
| + while (1) {
|
| + ValueType* entry = table + i;
|
| +
|
| + if (isEmptyBucket(*entry))
|
| + return LookupType(deletedEntry ? deletedEntry : entry, false);
|
| +
|
| + if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| + if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return LookupType(entry, true);
|
| +
|
| + if (isDeletedBucket(*entry))
|
| + deletedEntry = entry;
|
| + } else {
|
| + if (isDeletedBucket(*entry))
|
| + deletedEntry = entry;
|
| + else if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return LookupType(entry, true);
|
| }
|
| + UPDATE_PROBE_COUNTS();
|
| + if (!k)
|
| + k = 1 | doubleHash(h);
|
| + i = (i + k) & sizeMask;
|
| + }
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T>
|
| -inline typename HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::FullLookupType HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::fullLookupForWriting(const T& key)
|
| -{
|
| - ASSERT(!m_accessForbidden);
|
| - ASSERT(m_table);
|
| - registerModification();
|
| -
|
| - ValueType* table = m_table;
|
| - size_t k = 0;
|
| - size_t sizeMask = tableSizeMask();
|
| - unsigned h = HashTranslator::hash(key);
|
| - size_t i = h & sizeMask;
|
| -
|
| - UPDATE_ACCESS_COUNTS();
|
| -
|
| - ValueType* deletedEntry = nullptr;
|
| -
|
| - while (1) {
|
| - ValueType* entry = table + i;
|
| -
|
| - if (isEmptyBucket(*entry))
|
| - return makeLookupResult(deletedEntry ? deletedEntry : entry, false, h);
|
| -
|
| - if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| - if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return makeLookupResult(entry, true, h);
|
| -
|
| - if (isDeletedBucket(*entry))
|
| - deletedEntry = entry;
|
| - } else {
|
| - if (isDeletedBucket(*entry))
|
| - deletedEntry = entry;
|
| - else if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return makeLookupResult(entry, true, h);
|
| - }
|
| - UPDATE_PROBE_COUNTS();
|
| - if (!k)
|
| - k = 1 | doubleHash(h);
|
| - i = (i + k) & sizeMask;
|
| +inline typename HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::FullLookupType
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + fullLookupForWriting(const T& key) {
|
| + ASSERT(!m_accessForbidden);
|
| + ASSERT(m_table);
|
| + registerModification();
|
| +
|
| + ValueType* table = m_table;
|
| + size_t k = 0;
|
| + size_t sizeMask = tableSizeMask();
|
| + unsigned h = HashTranslator::hash(key);
|
| + size_t i = h & sizeMask;
|
| +
|
| + UPDATE_ACCESS_COUNTS();
|
| +
|
| + ValueType* deletedEntry = nullptr;
|
| +
|
| + while (1) {
|
| + ValueType* entry = table + i;
|
| +
|
| + if (isEmptyBucket(*entry))
|
| + return makeLookupResult(deletedEntry ? deletedEntry : entry, false, h);
|
| +
|
| + if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| + if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return makeLookupResult(entry, true, h);
|
| +
|
| + if (isDeletedBucket(*entry))
|
| + deletedEntry = entry;
|
| + } else {
|
| + if (isDeletedBucket(*entry))
|
| + deletedEntry = entry;
|
| + else if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return makeLookupResult(entry, true, h);
|
| }
|
| + UPDATE_PROBE_COUNTS();
|
| + if (!k)
|
| + k = 1 | doubleHash(h);
|
| + i = (i + k) & sizeMask;
|
| + }
|
| }
|
|
|
| -template <bool emptyValueIsZero> struct HashTableBucketInitializer;
|
| +template <bool emptyValueIsZero>
|
| +struct HashTableBucketInitializer;
|
|
|
| -template <> struct HashTableBucketInitializer<false> {
|
| - STATIC_ONLY(HashTableBucketInitializer);
|
| - template <typename Traits, typename Value> static void initialize(Value& bucket)
|
| - {
|
| - new (NotNull, &bucket) Value(Traits::emptyValue());
|
| - }
|
| +template <>
|
| +struct HashTableBucketInitializer<false> {
|
| + STATIC_ONLY(HashTableBucketInitializer);
|
| + template <typename Traits, typename Value>
|
| + static void initialize(Value& bucket) {
|
| + new (NotNull, &bucket) Value(Traits::emptyValue());
|
| + }
|
| };
|
|
|
| -template <> struct HashTableBucketInitializer<true> {
|
| - STATIC_ONLY(HashTableBucketInitializer);
|
| - template <typename Traits, typename Value> static void initialize(Value& bucket)
|
| - {
|
| - // This initializes the bucket without copying the empty value. That
|
| - // makes it possible to use this with types that don't support copying.
|
| - // The memset to 0 looks like a slow operation but is optimized by the
|
| - // compilers.
|
| - memset(&bucket, 0, sizeof(bucket));
|
| - }
|
| +template <>
|
| +struct HashTableBucketInitializer<true> {
|
| + STATIC_ONLY(HashTableBucketInitializer);
|
| + template <typename Traits, typename Value>
|
| + static void initialize(Value& bucket) {
|
| + // This initializes the bucket without copying the empty value. That
|
| + // makes it possible to use this with types that don't support copying.
|
| + // The memset to 0 looks like a slow operation but is optimized by the
|
| + // compilers.
|
| + memset(&bucket, 0, sizeof(bucket));
|
| + }
|
| };
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -inline void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::initializeBucket(ValueType& bucket)
|
| -{
|
| - HashTableBucketInitializer<Traits::emptyValueIsZero>::template initialize<Traits>(bucket);
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +inline void
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + initializeBucket(ValueType& bucket) {
|
| + HashTableBucketInitializer<Traits::emptyValueIsZero>::template initialize<
|
| + Traits>(bucket);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T, typename Extra>
|
| -typename HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::AddResult HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::add(const T& key, const Extra& extra)
|
| -{
|
| - ASSERT(!m_accessForbidden);
|
| - ASSERT(Allocator::isAllocationAllowed());
|
| - if (!m_table)
|
| - expand();
|
| -
|
| - ASSERT(m_table);
|
| -
|
| - ValueType* table = m_table;
|
| - size_t k = 0;
|
| - size_t sizeMask = tableSizeMask();
|
| - unsigned h = HashTranslator::hash(key);
|
| - size_t i = h & sizeMask;
|
| -
|
| - UPDATE_ACCESS_COUNTS();
|
| -
|
| - ValueType* deletedEntry = nullptr;
|
| - ValueType* entry;
|
| - while (1) {
|
| - entry = table + i;
|
| -
|
| - if (isEmptyBucket(*entry))
|
| - break;
|
| -
|
| - if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| - if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return AddResult(this, entry, false);
|
| -
|
| - if (isDeletedBucket(*entry))
|
| - deletedEntry = entry;
|
| - } else {
|
| - if (isDeletedBucket(*entry))
|
| - deletedEntry = entry;
|
| - else if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| - return AddResult(this, entry, false);
|
| - }
|
| - UPDATE_PROBE_COUNTS();
|
| - if (!k)
|
| - k = 1 | doubleHash(h);
|
| - i = (i + k) & sizeMask;
|
| +typename HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::AddResult
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + add(const T& key, const Extra& extra) {
|
| + ASSERT(!m_accessForbidden);
|
| + ASSERT(Allocator::isAllocationAllowed());
|
| + if (!m_table)
|
| + expand();
|
| +
|
| + ASSERT(m_table);
|
| +
|
| + ValueType* table = m_table;
|
| + size_t k = 0;
|
| + size_t sizeMask = tableSizeMask();
|
| + unsigned h = HashTranslator::hash(key);
|
| + size_t i = h & sizeMask;
|
| +
|
| + UPDATE_ACCESS_COUNTS();
|
| +
|
| + ValueType* deletedEntry = nullptr;
|
| + ValueType* entry;
|
| + while (1) {
|
| + entry = table + i;
|
| +
|
| + if (isEmptyBucket(*entry))
|
| + break;
|
| +
|
| + if (HashFunctions::safeToCompareToEmptyOrDeleted) {
|
| + if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return AddResult(this, entry, false);
|
| +
|
| + if (isDeletedBucket(*entry))
|
| + deletedEntry = entry;
|
| + } else {
|
| + if (isDeletedBucket(*entry))
|
| + deletedEntry = entry;
|
| + else if (HashTranslator::equal(Extractor::extract(*entry), key))
|
| + return AddResult(this, entry, false);
|
| }
|
| + UPDATE_PROBE_COUNTS();
|
| + if (!k)
|
| + k = 1 | doubleHash(h);
|
| + i = (i + k) & sizeMask;
|
| + }
|
|
|
| - registerModification();
|
| + registerModification();
|
|
|
| - if (deletedEntry) {
|
| - // Overwrite any data left over from last use, using placement new or
|
| - // memset.
|
| - initializeBucket(*deletedEntry);
|
| - entry = deletedEntry;
|
| - --m_deletedCount;
|
| - }
|
| + if (deletedEntry) {
|
| + // Overwrite any data left over from last use, using placement new or
|
| + // memset.
|
| + initializeBucket(*deletedEntry);
|
| + entry = deletedEntry;
|
| + --m_deletedCount;
|
| + }
|
|
|
| - HashTranslator::translate(*entry, key, extra);
|
| - ASSERT(!isEmptyOrDeletedBucket(*entry));
|
| + HashTranslator::translate(*entry, key, extra);
|
| + ASSERT(!isEmptyOrDeletedBucket(*entry));
|
|
|
| - ++m_keyCount;
|
| + ++m_keyCount;
|
|
|
| - if (shouldExpand())
|
| - entry = expand(entry);
|
| + if (shouldExpand())
|
| + entry = expand(entry);
|
|
|
| - return AddResult(this, entry, true);
|
| + return AddResult(this, entry, true);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T, typename Extra>
|
| -typename HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::AddResult HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::addPassingHashCode(const T& key, const Extra& extra)
|
| -{
|
| - ASSERT(!m_accessForbidden);
|
| - ASSERT(Allocator::isAllocationAllowed());
|
| - if (!m_table)
|
| - expand();
|
| -
|
| - FullLookupType lookupResult = fullLookupForWriting<HashTranslator>(key);
|
| -
|
| - ValueType* entry = lookupResult.first.first;
|
| - bool found = lookupResult.first.second;
|
| - unsigned h = lookupResult.second;
|
| -
|
| - if (found)
|
| - return AddResult(this, entry, false);
|
| -
|
| - registerModification();
|
| -
|
| - if (isDeletedBucket(*entry)) {
|
| - initializeBucket(*entry);
|
| - --m_deletedCount;
|
| - }
|
| -
|
| - HashTranslator::translate(*entry, key, extra, h);
|
| - ASSERT(!isEmptyOrDeletedBucket(*entry));
|
| -
|
| - ++m_keyCount;
|
| - if (shouldExpand())
|
| - entry = expand(entry);
|
| -
|
| - return AddResult(this, entry, true);
|
| +typename HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::AddResult
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + addPassingHashCode(const T& key, const Extra& extra) {
|
| + ASSERT(!m_accessForbidden);
|
| + ASSERT(Allocator::isAllocationAllowed());
|
| + if (!m_table)
|
| + expand();
|
| +
|
| + FullLookupType lookupResult = fullLookupForWriting<HashTranslator>(key);
|
| +
|
| + ValueType* entry = lookupResult.first.first;
|
| + bool found = lookupResult.first.second;
|
| + unsigned h = lookupResult.second;
|
| +
|
| + if (found)
|
| + return AddResult(this, entry, false);
|
| +
|
| + registerModification();
|
| +
|
| + if (isDeletedBucket(*entry)) {
|
| + initializeBucket(*entry);
|
| + --m_deletedCount;
|
| + }
|
| +
|
| + HashTranslator::translate(*entry, key, extra, h);
|
| + ASSERT(!isEmptyOrDeletedBucket(*entry));
|
| +
|
| + ++m_keyCount;
|
| + if (shouldExpand())
|
| + entry = expand(entry);
|
| +
|
| + return AddResult(this, entry, true);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::reinsert(ValueType& entry)
|
| -{
|
| - ASSERT(m_table);
|
| - registerModification();
|
| - ASSERT(!lookupForWriting(Extractor::extract(entry)).second);
|
| - ASSERT(!isDeletedBucket(*(lookupForWriting(Extractor::extract(entry)).first)));
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + reinsert(ValueType& entry) {
|
| + ASSERT(m_table);
|
| + registerModification();
|
| + ASSERT(!lookupForWriting(Extractor::extract(entry)).second);
|
| + ASSERT(
|
| + !isDeletedBucket(*(lookupForWriting(Extractor::extract(entry)).first)));
|
| #if DUMP_HASHTABLE_STATS
|
| - atomicIncrement(&HashTableStats::numReinserts);
|
| + atomicIncrement(&HashTableStats::numReinserts);
|
| #endif
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - ++m_stats->numReinserts;
|
| + ++m_stats->numReinserts;
|
| #endif
|
| - Value* newEntry = lookupForWriting(Extractor::extract(entry)).first;
|
| - Mover<ValueType, Allocator>::move(entry, *newEntry);
|
| + Value* newEntry = lookupForWriting(Extractor::extract(entry)).first;
|
| + Mover<ValueType, Allocator>::move(entry, *newEntry);
|
|
|
| - return newEntry;
|
| + return newEntry;
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T>
|
| -inline typename HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::iterator HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::find(const T& key)
|
| -{
|
| - ValueType* entry = lookup<HashTranslator>(key);
|
| - if (!entry)
|
| - return end();
|
| -
|
| - return makeKnownGoodIterator(entry);
|
| +inline typename HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::iterator
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + find(const T& key) {
|
| + ValueType* entry = lookup<HashTranslator>(key);
|
| + if (!entry)
|
| + return end();
|
| +
|
| + return makeKnownGoodIterator(entry);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T>
|
| -inline typename HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::const_iterator HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::find(const T& key) const
|
| -{
|
| - ValueType* entry = const_cast<HashTable*>(this)->lookup<HashTranslator>(key);
|
| - if (!entry)
|
| - return end();
|
| -
|
| - return makeKnownGoodConstIterator(entry);
|
| +inline typename HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::const_iterator
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + find(const T& key) const {
|
| + ValueType* entry = const_cast<HashTable*>(this)->lookup<HashTranslator>(key);
|
| + if (!entry)
|
| + return end();
|
| +
|
| + return makeKnownGoodConstIterator(entry);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename HashTranslator, typename T>
|
| -bool HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::contains(const T& key) const
|
| -{
|
| - return const_cast<HashTable*>(this)->lookup<HashTranslator>(key);
|
| +bool HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::contains(const T& key) const {
|
| + return const_cast<HashTable*>(this)->lookup<HashTranslator>(key);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::remove(ValueType* pos)
|
| -{
|
| - registerModification();
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +void HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::remove(ValueType* pos) {
|
| + registerModification();
|
| #if DUMP_HASHTABLE_STATS
|
| - atomicIncrement(&HashTableStats::numRemoves);
|
| + atomicIncrement(&HashTableStats::numRemoves);
|
| #endif
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - ++m_stats->numRemoves;
|
| + ++m_stats->numRemoves;
|
| #endif
|
|
|
| - ASSERT(!m_accessForbidden);
|
| + ASSERT(!m_accessForbidden);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = true;
|
| + m_accessForbidden = true;
|
| #endif
|
| - deleteBucket(*pos);
|
| + deleteBucket(*pos);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = false;
|
| + m_accessForbidden = false;
|
| #endif
|
| - ++m_deletedCount;
|
| - --m_keyCount;
|
| + ++m_deletedCount;
|
| + --m_keyCount;
|
|
|
| - if (shouldShrink())
|
| - shrink();
|
| + if (shouldShrink())
|
| + shrink();
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -inline void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::remove(iterator it)
|
| -{
|
| - if (it == end())
|
| - return;
|
| - remove(const_cast<ValueType*>(it.m_iterator.m_position));
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +inline void
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + remove(iterator it) {
|
| + if (it == end())
|
| + return;
|
| + remove(const_cast<ValueType*>(it.m_iterator.m_position));
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -inline void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::remove(const_iterator it)
|
| -{
|
| - if (it == end())
|
| - return;
|
| - remove(const_cast<ValueType*>(it.m_position));
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +inline void
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + remove(const_iterator it) {
|
| + if (it == end())
|
| + return;
|
| + remove(const_cast<ValueType*>(it.m_position));
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -inline void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::remove(KeyPeekInType key)
|
| -{
|
| - remove(find(key));
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +inline void
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + remove(KeyPeekInType key) {
|
| + remove(find(key));
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::allocateTable(unsigned size)
|
| -{
|
| - size_t allocSize = size * sizeof(ValueType);
|
| - ValueType* result;
|
| - // Assert that we will not use memset on things with a vtable entry. The
|
| - // compiler will also check this on some platforms. We would like to check
|
| - // this on the whole value (key-value pair), but std::is_polymorphic will return
|
| - // false for a pair of two types, even if one of the components is
|
| - // polymorphic.
|
| - static_assert(!Traits::emptyValueIsZero || !std::is_polymorphic<KeyType>::value, "empty value cannot be zero for things with a vtable");
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + allocateTable(unsigned size) {
|
| + size_t allocSize = size * sizeof(ValueType);
|
| + ValueType* result;
|
| + // Assert that we will not use memset on things with a vtable entry. The
|
| + // compiler will also check this on some platforms. We would like to check
|
| + // this on the whole value (key-value pair), but std::is_polymorphic will return
|
| + // false for a pair of two types, even if one of the components is
|
| + // polymorphic.
|
| + static_assert(
|
| + !Traits::emptyValueIsZero || !std::is_polymorphic<KeyType>::value,
|
| + "empty value cannot be zero for things with a vtable");
|
|
|
| #if ENABLE(OILPAN)
|
| - static_assert(Allocator::isGarbageCollected
|
| - || ((!AllowsOnlyPlacementNew<KeyType>::value || !NeedsTracing<KeyType>::value)
|
| - && (!AllowsOnlyPlacementNew<ValueType>::value || !NeedsTracing<ValueType>::value))
|
| - , "Cannot put DISALLOW_NEW_EXCEPT_PLACEMENT_NEW objects that have trace methods into an off-heap HashTable");
|
| + static_assert(Allocator::isGarbageCollected ||
|
| + ((!AllowsOnlyPlacementNew<KeyType>::value ||
|
| + !NeedsTracing<KeyType>::value) &&
|
| + (!AllowsOnlyPlacementNew<ValueType>::value ||
|
| + !NeedsTracing<ValueType>::value)),
|
| + "Cannot put DISALLOW_NEW_EXCEPT_PLACEMENT_NEW objects that "
|
| + "have trace methods into an off-heap HashTable");
|
| #endif
|
| - if (Traits::emptyValueIsZero) {
|
| - result = Allocator::template allocateZeroedHashTableBacking<ValueType, HashTable>(allocSize);
|
| - } else {
|
| - result = Allocator::template allocateHashTableBacking<ValueType, HashTable>(allocSize);
|
| - for (unsigned i = 0; i < size; i++)
|
| - initializeBucket(result[i]);
|
| - }
|
| - return result;
|
| + if (Traits::emptyValueIsZero) {
|
| + result = Allocator::template allocateZeroedHashTableBacking<ValueType,
|
| + HashTable>(
|
| + allocSize);
|
| + } else {
|
| + result = Allocator::template allocateHashTableBacking<ValueType, HashTable>(
|
| + allocSize);
|
| + for (unsigned i = 0; i < size; i++)
|
| + initializeBucket(result[i]);
|
| + }
|
| + return result;
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::deleteAllBucketsAndDeallocate(ValueType* table, unsigned size)
|
| -{
|
| - if (!IsTriviallyDestructible<ValueType>::value) {
|
| - for (unsigned i = 0; i < size; ++i) {
|
| - // This code is called when the hash table is cleared or resized. We
|
| - // have allocated a new backing store and we need to run the
|
| - // destructors on the old backing store, as it is being freed. If we
|
| - // are GCing we need to both call the destructor and mark the bucket
|
| - // as deleted, otherwise the destructor gets called again when the
|
| - // GC finds the backing store. With the default allocator it's
|
| - // enough to call the destructor, since we will free the memory
|
| - // explicitly and we won't see the memory with the bucket again.
|
| - if (Allocator::isGarbageCollected) {
|
| - if (!isEmptyOrDeletedBucket(table[i]))
|
| - deleteBucket(table[i]);
|
| - } else {
|
| - if (!isDeletedBucket(table[i]))
|
| - table[i].~ValueType();
|
| - }
|
| - }
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +void HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::deleteAllBucketsAndDeallocate(ValueType* table,
|
| + unsigned size) {
|
| + if (!IsTriviallyDestructible<ValueType>::value) {
|
| + for (unsigned i = 0; i < size; ++i) {
|
| + // This code is called when the hash table is cleared or resized. We
|
| + // have allocated a new backing store and we need to run the
|
| + // destructors on the old backing store, as it is being freed. If we
|
| + // are GCing we need to both call the destructor and mark the bucket
|
| + // as deleted, otherwise the destructor gets called again when the
|
| + // GC finds the backing store. With the default allocator it's
|
| + // enough to call the destructor, since we will free the memory
|
| + // explicitly and we won't see the memory with the bucket again.
|
| + if (Allocator::isGarbageCollected) {
|
| + if (!isEmptyOrDeletedBucket(table[i]))
|
| + deleteBucket(table[i]);
|
| + } else {
|
| + if (!isDeletedBucket(table[i]))
|
| + table[i].~ValueType();
|
| + }
|
| }
|
| - Allocator::freeHashTableBacking(table);
|
| + }
|
| + Allocator::freeHashTableBacking(table);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::expand(Value* entry)
|
| -{
|
| - unsigned newSize;
|
| - if (!m_tableSize) {
|
| - newSize = KeyTraits::minimumTableSize;
|
| - } else if (mustRehashInPlace()) {
|
| - newSize = m_tableSize;
|
| - } else {
|
| - newSize = m_tableSize * 2;
|
| - RELEASE_ASSERT(newSize > m_tableSize);
|
| - }
|
| -
|
| - return rehash(newSize, entry);
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + expand(Value* entry) {
|
| + unsigned newSize;
|
| + if (!m_tableSize) {
|
| + newSize = KeyTraits::minimumTableSize;
|
| + } else if (mustRehashInPlace()) {
|
| + newSize = m_tableSize;
|
| + } else {
|
| + newSize = m_tableSize * 2;
|
| + RELEASE_ASSERT(newSize > m_tableSize);
|
| + }
|
| +
|
| + return rehash(newSize, entry);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::expandBuffer(unsigned newTableSize, Value* entry, bool& success)
|
| -{
|
| - success = false;
|
| - ASSERT(m_tableSize < newTableSize);
|
| - if (!Allocator::expandHashTableBacking(m_table, newTableSize * sizeof(ValueType)))
|
| - return nullptr;
|
| -
|
| - success = true;
|
| -
|
| - Value* newEntry = nullptr;
|
| - unsigned oldTableSize = m_tableSize;
|
| - ValueType* originalTable = m_table;
|
| -
|
| - ValueType* temporaryTable = allocateTable(oldTableSize);
|
| - for (unsigned i = 0; i < oldTableSize; i++) {
|
| - if (&m_table[i] == entry)
|
| - newEntry = &temporaryTable[i];
|
| - if (isEmptyOrDeletedBucket(m_table[i])) {
|
| - ASSERT(&m_table[i] != entry);
|
| - if (Traits::emptyValueIsZero) {
|
| - memset(&temporaryTable[i], 0, sizeof(ValueType));
|
| - } else {
|
| - initializeBucket(temporaryTable[i]);
|
| - }
|
| - } else {
|
| - Mover<ValueType, Allocator>::move(m_table[i], temporaryTable[i]);
|
| - }
|
| - }
|
| - m_table = temporaryTable;
|
| -
|
| - if (Traits::emptyValueIsZero) {
|
| - memset(originalTable, 0, newTableSize * sizeof(ValueType));
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + expandBuffer(unsigned newTableSize, Value* entry, bool& success) {
|
| + success = false;
|
| + ASSERT(m_tableSize < newTableSize);
|
| + if (!Allocator::expandHashTableBacking(m_table,
|
| + newTableSize * sizeof(ValueType)))
|
| + return nullptr;
|
| +
|
| + success = true;
|
| +
|
| + Value* newEntry = nullptr;
|
| + unsigned oldTableSize = m_tableSize;
|
| + ValueType* originalTable = m_table;
|
| +
|
| + ValueType* temporaryTable = allocateTable(oldTableSize);
|
| + for (unsigned i = 0; i < oldTableSize; i++) {
|
| + if (&m_table[i] == entry)
|
| + newEntry = &temporaryTable[i];
|
| + if (isEmptyOrDeletedBucket(m_table[i])) {
|
| + ASSERT(&m_table[i] != entry);
|
| + if (Traits::emptyValueIsZero) {
|
| + memset(&temporaryTable[i], 0, sizeof(ValueType));
|
| + } else {
|
| + initializeBucket(temporaryTable[i]);
|
| + }
|
| } else {
|
| - for (unsigned i = 0; i < newTableSize; i++)
|
| - initializeBucket(originalTable[i]);
|
| + Mover<ValueType, Allocator>::move(m_table[i], temporaryTable[i]);
|
| }
|
| - newEntry = rehashTo(originalTable, newTableSize, newEntry);
|
| -
|
| - ASSERT(!m_accessForbidden);
|
| + }
|
| + m_table = temporaryTable;
|
| +
|
| + if (Traits::emptyValueIsZero) {
|
| + memset(originalTable, 0, newTableSize * sizeof(ValueType));
|
| + } else {
|
| + for (unsigned i = 0; i < newTableSize; i++)
|
| + initializeBucket(originalTable[i]);
|
| + }
|
| + newEntry = rehashTo(originalTable, newTableSize, newEntry);
|
| +
|
| + ASSERT(!m_accessForbidden);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = true;
|
| + m_accessForbidden = true;
|
| #endif
|
| - deleteAllBucketsAndDeallocate(temporaryTable, oldTableSize);
|
| + deleteAllBucketsAndDeallocate(temporaryTable, oldTableSize);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = false;
|
| + m_accessForbidden = false;
|
| #endif
|
|
|
| - return newEntry;
|
| + return newEntry;
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::rehashTo(ValueType* newTable, unsigned newTableSize, Value* entry)
|
| -{
|
| - unsigned oldTableSize = m_tableSize;
|
| - ValueType* oldTable = m_table;
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + rehashTo(ValueType* newTable, unsigned newTableSize, Value* entry) {
|
| + unsigned oldTableSize = m_tableSize;
|
| + ValueType* oldTable = m_table;
|
|
|
| #if DUMP_HASHTABLE_STATS
|
| - if (oldTableSize != 0)
|
| - atomicIncrement(&HashTableStats::numRehashes);
|
| + if (oldTableSize != 0)
|
| + atomicIncrement(&HashTableStats::numRehashes);
|
| #endif
|
|
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - if (oldTableSize != 0)
|
| - ++m_stats->numRehashes;
|
| + if (oldTableSize != 0)
|
| + ++m_stats->numRehashes;
|
| #endif
|
|
|
| - m_table = newTable;
|
| - m_tableSize = newTableSize;
|
| + m_table = newTable;
|
| + m_tableSize = newTableSize;
|
|
|
| - Value* newEntry = nullptr;
|
| - for (unsigned i = 0; i != oldTableSize; ++i) {
|
| - if (isEmptyOrDeletedBucket(oldTable[i])) {
|
| - ASSERT(&oldTable[i] != entry);
|
| - continue;
|
| - }
|
| - Value* reinsertedEntry = reinsert(oldTable[i]);
|
| - if (&oldTable[i] == entry) {
|
| - ASSERT(!newEntry);
|
| - newEntry = reinsertedEntry;
|
| - }
|
| + Value* newEntry = nullptr;
|
| + for (unsigned i = 0; i != oldTableSize; ++i) {
|
| + if (isEmptyOrDeletedBucket(oldTable[i])) {
|
| + ASSERT(&oldTable[i] != entry);
|
| + continue;
|
| + }
|
| + Value* reinsertedEntry = reinsert(oldTable[i]);
|
| + if (&oldTable[i] == entry) {
|
| + ASSERT(!newEntry);
|
| + newEntry = reinsertedEntry;
|
| }
|
| + }
|
|
|
| - m_deletedCount = 0;
|
| + m_deletedCount = 0;
|
|
|
| - return newEntry;
|
| + return newEntry;
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -Value* HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::rehash(unsigned newTableSize, Value* entry)
|
| -{
|
| - unsigned oldTableSize = m_tableSize;
|
| - ValueType* oldTable = m_table;
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +Value*
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + rehash(unsigned newTableSize, Value* entry) {
|
| + unsigned oldTableSize = m_tableSize;
|
| + ValueType* oldTable = m_table;
|
|
|
| #if DUMP_HASHTABLE_STATS
|
| - if (oldTableSize != 0)
|
| - atomicIncrement(&HashTableStats::numRehashes);
|
| + if (oldTableSize != 0)
|
| + atomicIncrement(&HashTableStats::numRehashes);
|
| #endif
|
|
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - if (oldTableSize != 0)
|
| - ++m_stats->numRehashes;
|
| + if (oldTableSize != 0)
|
| + ++m_stats->numRehashes;
|
| #endif
|
|
|
| - // The Allocator::isGarbageCollected check is not needed. The check is just
|
| - // a static hint for a compiler to indicate that Base::expandBuffer returns
|
| - // false if Allocator is a PartitionAllocator.
|
| - if (Allocator::isGarbageCollected && newTableSize > oldTableSize) {
|
| - bool success;
|
| - Value* newEntry = expandBuffer(newTableSize, entry, success);
|
| - if (success)
|
| - return newEntry;
|
| - }
|
| + // The Allocator::isGarbageCollected check is not needed. The check is just
|
| + // a static hint for a compiler to indicate that Base::expandBuffer returns
|
| + // false if Allocator is a PartitionAllocator.
|
| + if (Allocator::isGarbageCollected && newTableSize > oldTableSize) {
|
| + bool success;
|
| + Value* newEntry = expandBuffer(newTableSize, entry, success);
|
| + if (success)
|
| + return newEntry;
|
| + }
|
|
|
| - ValueType* newTable = allocateTable(newTableSize);
|
| - Value* newEntry = rehashTo(newTable, newTableSize, entry);
|
| + ValueType* newTable = allocateTable(newTableSize);
|
| + Value* newEntry = rehashTo(newTable, newTableSize, entry);
|
|
|
| - ASSERT(!m_accessForbidden);
|
| + ASSERT(!m_accessForbidden);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = true;
|
| + m_accessForbidden = true;
|
| #endif
|
| - deleteAllBucketsAndDeallocate(oldTable, oldTableSize);
|
| + deleteAllBucketsAndDeallocate(oldTable, oldTableSize);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = false;
|
| + m_accessForbidden = false;
|
| #endif
|
|
|
| - return newEntry;
|
| + return newEntry;
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::clear()
|
| -{
|
| - registerModification();
|
| - if (!m_table)
|
| - return;
|
| -
|
| - ASSERT(!m_accessForbidden);
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +void HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::clear() {
|
| + registerModification();
|
| + if (!m_table)
|
| + return;
|
| +
|
| + ASSERT(!m_accessForbidden);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = true;
|
| + m_accessForbidden = true;
|
| #endif
|
| - deleteAllBucketsAndDeallocate(m_table, m_tableSize);
|
| + deleteAllBucketsAndDeallocate(m_table, m_tableSize);
|
| #if ENABLE(ASSERT)
|
| - m_accessForbidden = false;
|
| + m_accessForbidden = false;
|
| #endif
|
| - m_table = nullptr;
|
| - m_tableSize = 0;
|
| - m_keyCount = 0;
|
| + m_table = nullptr;
|
| + m_tableSize = 0;
|
| + m_keyCount = 0;
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::HashTable(const HashTable& other)
|
| - : m_table(nullptr)
|
| - , m_tableSize(0)
|
| - , m_keyCount(0)
|
| - , m_deletedCount(0)
|
| - , m_queueFlag(false)
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| + HashTable(const HashTable& other)
|
| + : m_table(nullptr),
|
| + m_tableSize(0),
|
| + m_keyCount(0),
|
| + m_deletedCount(0),
|
| + m_queueFlag(false)
|
| #if ENABLE(ASSERT)
|
| - , m_accessForbidden(false)
|
| - , m_modifications(0)
|
| + ,
|
| + m_accessForbidden(false),
|
| + m_modifications(0)
|
| #endif
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - , m_stats(adoptPtr(new Stats(*other.m_stats)))
|
| + ,
|
| + m_stats(adoptPtr(new Stats(*other.m_stats)))
|
| #endif
|
| {
|
| - // Copy the hash table the dumb way, by adding each element to the new
|
| - // table. It might be more efficient to copy the table slots, but it's not
|
| - // clear that efficiency is needed.
|
| - const_iterator end = other.end();
|
| - for (const_iterator it = other.begin(); it != end; ++it)
|
| - add(*it);
|
| + // Copy the hash table the dumb way, by adding each element to the new
|
| + // table. It might be more efficient to copy the table slots, but it's not
|
| + // clear that efficiency is needed.
|
| + const_iterator end = other.end();
|
| + for (const_iterator it = other.begin(); it != end; ++it)
|
| + add(*it);
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::swap(HashTable& other)
|
| -{
|
| - ASSERT(!m_accessForbidden);
|
| - std::swap(m_table, other.m_table);
|
| - std::swap(m_tableSize, other.m_tableSize);
|
| - std::swap(m_keyCount, other.m_keyCount);
|
| - // std::swap does not work for bit fields.
|
| - unsigned deleted = m_deletedCount;
|
| - m_deletedCount = other.m_deletedCount;
|
| - other.m_deletedCount = deleted;
|
| - ASSERT(!m_queueFlag);
|
| - ASSERT(!other.m_queueFlag);
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +void HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::swap(HashTable& other) {
|
| + ASSERT(!m_accessForbidden);
|
| + std::swap(m_table, other.m_table);
|
| + std::swap(m_tableSize, other.m_tableSize);
|
| + std::swap(m_keyCount, other.m_keyCount);
|
| + // std::swap does not work for bit fields.
|
| + unsigned deleted = m_deletedCount;
|
| + m_deletedCount = other.m_deletedCount;
|
| + other.m_deletedCount = deleted;
|
| + ASSERT(!m_queueFlag);
|
| + ASSERT(!other.m_queueFlag);
|
|
|
| #if ENABLE(ASSERT)
|
| - std::swap(m_modifications, other.m_modifications);
|
| + std::swap(m_modifications, other.m_modifications);
|
| #endif
|
|
|
| #if DUMP_HASHTABLE_STATS_PER_TABLE
|
| - m_stats.swap(other.m_stats);
|
| + m_stats.swap(other.m_stats);
|
| #endif
|
| }
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>& HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::operator=(const HashTable& other)
|
| -{
|
| - HashTable tmp(other);
|
| - swap(tmp);
|
| - return *this;
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>&
|
| +HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::
|
| +operator=(const HashTable& other) {
|
| + HashTable tmp(other);
|
| + swap(tmp);
|
| + return *this;
|
| }
|
|
|
| -template <WeakHandlingFlag weakHandlingFlag, typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <WeakHandlingFlag weakHandlingFlag,
|
| + typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| struct WeakProcessingHashTableHelper;
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -struct WeakProcessingHashTableHelper<NoWeakHandlingInCollections, Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> {
|
| - STATIC_ONLY(WeakProcessingHashTableHelper);
|
| - static void process(typename Allocator::Visitor* visitor, void* closure) {}
|
| - static void ephemeronIteration(typename Allocator::Visitor* visitor, void* closure) {}
|
| - static void ephemeronIterationDone(typename Allocator::Visitor* visitor, void* closure) {}
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +struct WeakProcessingHashTableHelper<NoWeakHandlingInCollections,
|
| + Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator> {
|
| + STATIC_ONLY(WeakProcessingHashTableHelper);
|
| + static void process(typename Allocator::Visitor* visitor, void* closure) {}
|
| + static void ephemeronIteration(typename Allocator::Visitor* visitor,
|
| + void* closure) {}
|
| + static void ephemeronIterationDone(typename Allocator::Visitor* visitor,
|
| + void* closure) {}
|
| };
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| -struct WeakProcessingHashTableHelper<WeakHandlingInCollections, Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> {
|
| - STATIC_ONLY(WeakProcessingHashTableHelper);
|
| - // Used for purely weak and for weak-and-strong tables (ephemerons).
|
| - static void process(typename Allocator::Visitor* visitor, void* closure)
|
| - {
|
| - typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> HashTableType;
|
| - HashTableType* table = reinterpret_cast<HashTableType*>(closure);
|
| - if (!table->m_table)
|
| - return;
|
| - // Now perform weak processing (this is a no-op if the backing was
|
| - // accessible through an iterator and was already marked strongly).
|
| - typedef typename HashTableType::ValueType ValueType;
|
| - for (ValueType* element = table->m_table + table->m_tableSize - 1; element >= table->m_table; element--) {
|
| - if (!HashTableType::isEmptyOrDeletedBucket(*element)) {
|
| - // At this stage calling trace can make no difference
|
| - // (everything is already traced), but we use the return value
|
| - // to remove things from the collection.
|
| -
|
| - // FIXME: This should be rewritten so that this can check if the
|
| - // element is dead without calling trace, which is semantically
|
| - // not correct to be called in weak processing stage.
|
| - if (TraceInCollectionTrait<WeakHandlingInCollections, WeakPointersActWeak, ValueType, Traits>::trace(visitor, *element)) {
|
| - table->registerModification();
|
| - HashTableType::deleteBucket(*element); // Also calls the destructor.
|
| - table->m_deletedCount++;
|
| - table->m_keyCount--;
|
| - // We don't rehash the backing until the next add or delete,
|
| - // because that would cause allocation during GC.
|
| - }
|
| - }
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| +struct WeakProcessingHashTableHelper<WeakHandlingInCollections,
|
| + Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator> {
|
| + STATIC_ONLY(WeakProcessingHashTableHelper);
|
| + // Used for purely weak and for weak-and-strong tables (ephemerons).
|
| + static void process(typename Allocator::Visitor* visitor, void* closure) {
|
| + typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits,
|
| + Allocator>
|
| + HashTableType;
|
| + HashTableType* table = reinterpret_cast<HashTableType*>(closure);
|
| + if (!table->m_table)
|
| + return;
|
| + // Now perform weak processing (this is a no-op if the backing was
|
| + // accessible through an iterator and was already marked strongly).
|
| + typedef typename HashTableType::ValueType ValueType;
|
| + for (ValueType* element = table->m_table + table->m_tableSize - 1;
|
| + element >= table->m_table; element--) {
|
| + if (!HashTableType::isEmptyOrDeletedBucket(*element)) {
|
| + // At this stage calling trace can make no difference
|
| + // (everything is already traced), but we use the return value
|
| + // to remove things from the collection.
|
| +
|
| + // FIXME: This should be rewritten so that this can check if the
|
| + // element is dead without calling trace, which is semantically
|
| + // not correct to be called in weak processing stage.
|
| + if (TraceInCollectionTrait<WeakHandlingInCollections,
|
| + WeakPointersActWeak, ValueType,
|
| + Traits>::trace(visitor, *element)) {
|
| + table->registerModification();
|
| + HashTableType::deleteBucket(*element); // Also calls the destructor.
|
| + table->m_deletedCount++;
|
| + table->m_keyCount--;
|
| + // We don't rehash the backing until the next add or delete,
|
| + // because that would cause allocation during GC.
|
| }
|
| + }
|
| }
|
| -
|
| - // Called repeatedly for tables that have both weak and strong pointers.
|
| - static void ephemeronIteration(typename Allocator::Visitor* visitor, void* closure)
|
| - {
|
| - typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> HashTableType;
|
| - HashTableType* table = reinterpret_cast<HashTableType*>(closure);
|
| - ASSERT(table->m_table);
|
| - // Check the hash table for elements that we now know will not be
|
| - // removed by weak processing. Those elements need to have their strong
|
| - // pointers traced.
|
| - typedef typename HashTableType::ValueType ValueType;
|
| - for (ValueType* element = table->m_table + table->m_tableSize - 1; element >= table->m_table; element--) {
|
| - if (!HashTableType::isEmptyOrDeletedBucket(*element))
|
| - TraceInCollectionTrait<WeakHandlingInCollections, WeakPointersActWeak, ValueType, Traits>::trace(visitor, *element);
|
| - }
|
| - }
|
| -
|
| - // Called when the ephemeron iteration is done and before running the per
|
| - // thread weak processing. It is guaranteed to be called before any thread
|
| - // is resumed.
|
| - static void ephemeronIterationDone(typename Allocator::Visitor* visitor, void* closure)
|
| - {
|
| - typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator> HashTableType;
|
| - HashTableType* table = reinterpret_cast<HashTableType*>(closure);
|
| - ASSERT(Allocator::weakTableRegistered(visitor, table));
|
| - table->clearEnqueued();
|
| + }
|
| +
|
| + // Called repeatedly for tables that have both weak and strong pointers.
|
| + static void ephemeronIteration(typename Allocator::Visitor* visitor,
|
| + void* closure) {
|
| + typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits,
|
| + Allocator>
|
| + HashTableType;
|
| + HashTableType* table = reinterpret_cast<HashTableType*>(closure);
|
| + ASSERT(table->m_table);
|
| + // Check the hash table for elements that we now know will not be
|
| + // removed by weak processing. Those elements need to have their strong
|
| + // pointers traced.
|
| + typedef typename HashTableType::ValueType ValueType;
|
| + for (ValueType* element = table->m_table + table->m_tableSize - 1;
|
| + element >= table->m_table; element--) {
|
| + if (!HashTableType::isEmptyOrDeletedBucket(*element))
|
| + TraceInCollectionTrait<WeakHandlingInCollections, WeakPointersActWeak,
|
| + ValueType, Traits>::trace(visitor, *element);
|
| }
|
| + }
|
| +
|
| + // Called when the ephemeron iteration is done and before running the per
|
| + // thread weak processing. It is guaranteed to be called before any thread
|
| + // is resumed.
|
| + static void ephemeronIterationDone(typename Allocator::Visitor* visitor,
|
| + void* closure) {
|
| + typedef HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits,
|
| + Allocator>
|
| + HashTableType;
|
| + HashTableType* table = reinterpret_cast<HashTableType*>(closure);
|
| + ASSERT(Allocator::weakTableRegistered(visitor, table));
|
| + table->clearEnqueued();
|
| + }
|
| };
|
|
|
| -template <typename Key, typename Value, typename Extractor, typename HashFunctions, typename Traits, typename KeyTraits, typename Allocator>
|
| +template <typename Key,
|
| + typename Value,
|
| + typename Extractor,
|
| + typename HashFunctions,
|
| + typename Traits,
|
| + typename KeyTraits,
|
| + typename Allocator>
|
| template <typename VisitorDispatcher>
|
| -void HashTable<Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::trace(VisitorDispatcher visitor)
|
| -{
|
| - // If someone else already marked the backing and queued up the trace and/or
|
| - // weak callback then we are done. This optimization does not happen for
|
| - // ListHashSet since its iterator does not point at the backing.
|
| - if (!m_table || Allocator::isHeapObjectAlive(m_table))
|
| - return;
|
| - // Normally, we mark the backing store without performing trace. This means
|
| - // it is marked live, but the pointers inside it are not marked. Instead we
|
| - // will mark the pointers below. However, for backing stores that contain
|
| - // weak pointers the handling is rather different. We don't mark the
|
| - // backing store here, so the marking GC will leave the backing unmarked. If
|
| - // the backing is found in any other way than through its HashTable (ie from
|
| - // an iterator) then the mark bit will be set and the pointers will be
|
| - // marked strongly, avoiding problems with iterating over things that
|
| - // disappear due to weak processing while we are iterating over them. We
|
| - // register the backing store pointer for delayed marking which will take
|
| - // place after we know if the backing is reachable from elsewhere. We also
|
| - // register a weakProcessing callback which will perform weak processing if
|
| - // needed.
|
| - if (Traits::weakHandlingFlag == NoWeakHandlingInCollections) {
|
| - Allocator::markNoTracing(visitor, m_table);
|
| - } else {
|
| - Allocator::registerDelayedMarkNoTracing(visitor, m_table);
|
| - // Since we're delaying marking this HashTable, it is possible that the
|
| - // registerWeakMembers is called multiple times (in rare
|
| - // cases). However, it shouldn't cause any issue.
|
| - Allocator::registerWeakMembers(visitor, this, m_table, WeakProcessingHashTableHelper<Traits::weakHandlingFlag, Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::process);
|
| +void HashTable<Key,
|
| + Value,
|
| + Extractor,
|
| + HashFunctions,
|
| + Traits,
|
| + KeyTraits,
|
| + Allocator>::trace(VisitorDispatcher visitor) {
|
| + // If someone else already marked the backing and queued up the trace and/or
|
| + // weak callback then we are done. This optimization does not happen for
|
| + // ListHashSet since its iterator does not point at the backing.
|
| + if (!m_table || Allocator::isHeapObjectAlive(m_table))
|
| + return;
|
| + // Normally, we mark the backing store without performing trace. This means
|
| + // it is marked live, but the pointers inside it are not marked. Instead we
|
| + // will mark the pointers below. However, for backing stores that contain
|
| + // weak pointers the handling is rather different. We don't mark the
|
| + // backing store here, so the marking GC will leave the backing unmarked. If
|
| + // the backing is found in any other way than through its HashTable (ie from
|
| + // an iterator) then the mark bit will be set and the pointers will be
|
| + // marked strongly, avoiding problems with iterating over things that
|
| + // disappear due to weak processing while we are iterating over them. We
|
| + // register the backing store pointer for delayed marking which will take
|
| + // place after we know if the backing is reachable from elsewhere. We also
|
| + // register a weakProcessing callback which will perform weak processing if
|
| + // needed.
|
| + if (Traits::weakHandlingFlag == NoWeakHandlingInCollections) {
|
| + Allocator::markNoTracing(visitor, m_table);
|
| + } else {
|
| + Allocator::registerDelayedMarkNoTracing(visitor, m_table);
|
| + // Since we're delaying marking this HashTable, it is possible that the
|
| + // registerWeakMembers is called multiple times (in rare
|
| + // cases). However, it shouldn't cause any issue.
|
| + Allocator::registerWeakMembers(
|
| + visitor, this, m_table,
|
| + WeakProcessingHashTableHelper<Traits::weakHandlingFlag, Key, Value,
|
| + Extractor, HashFunctions, Traits,
|
| + KeyTraits, Allocator>::process);
|
| + }
|
| + if (NeedsTracingTrait<Traits>::value) {
|
| + if (Traits::weakHandlingFlag == WeakHandlingInCollections) {
|
| + // If we have both strong and weak pointers in the collection then
|
| + // we queue up the collection for fixed point iteration a la
|
| + // Ephemerons:
|
| + // http://dl.acm.org/citation.cfm?doid=263698.263733 - see also
|
| + // http://www.jucs.org/jucs_14_21/eliminating_cycles_in_weak
|
| + ASSERT(!enqueued() || Allocator::weakTableRegistered(visitor, this));
|
| + if (!enqueued()) {
|
| + Allocator::registerWeakTable(
|
| + visitor, this,
|
| + WeakProcessingHashTableHelper<
|
| + Traits::weakHandlingFlag, Key, Value, Extractor, HashFunctions,
|
| + Traits, KeyTraits, Allocator>::ephemeronIteration,
|
| + WeakProcessingHashTableHelper<
|
| + Traits::weakHandlingFlag, Key, Value, Extractor, HashFunctions,
|
| + Traits, KeyTraits, Allocator>::ephemeronIterationDone);
|
| + setEnqueued();
|
| + }
|
| + // We don't need to trace the elements here, since registering as a
|
| + // weak table above will cause them to be traced (perhaps several
|
| + // times). It's better to wait until everything else is traced
|
| + // before tracing the elements for the first time; this may reduce
|
| + // (by one) the number of iterations needed to get to a fixed point.
|
| + return;
|
| }
|
| - if (NeedsTracingTrait<Traits>::value) {
|
| - if (Traits::weakHandlingFlag == WeakHandlingInCollections) {
|
| - // If we have both strong and weak pointers in the collection then
|
| - // we queue up the collection for fixed point iteration a la
|
| - // Ephemerons:
|
| - // http://dl.acm.org/citation.cfm?doid=263698.263733 - see also
|
| - // http://www.jucs.org/jucs_14_21/eliminating_cycles_in_weak
|
| - ASSERT(!enqueued() || Allocator::weakTableRegistered(visitor, this));
|
| - if (!enqueued()) {
|
| - Allocator::registerWeakTable(visitor, this,
|
| - WeakProcessingHashTableHelper<Traits::weakHandlingFlag, Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::ephemeronIteration,
|
| - WeakProcessingHashTableHelper<Traits::weakHandlingFlag, Key, Value, Extractor, HashFunctions, Traits, KeyTraits, Allocator>::ephemeronIterationDone);
|
| - setEnqueued();
|
| - }
|
| - // We don't need to trace the elements here, since registering as a
|
| - // weak table above will cause them to be traced (perhaps several
|
| - // times). It's better to wait until everything else is traced
|
| - // before tracing the elements for the first time; this may reduce
|
| - // (by one) the number of iterations needed to get to a fixed point.
|
| - return;
|
| - }
|
| - for (ValueType* element = m_table + m_tableSize - 1; element >= m_table; element--) {
|
| - if (!isEmptyOrDeletedBucket(*element))
|
| - Allocator::template trace<VisitorDispatcher, ValueType, Traits>(visitor, *element);
|
| - }
|
| + for (ValueType* element = m_table + m_tableSize - 1; element >= m_table;
|
| + element--) {
|
| + if (!isEmptyOrDeletedBucket(*element))
|
| + Allocator::template trace<VisitorDispatcher, ValueType, Traits>(
|
| + visitor, *element);
|
| }
|
| + }
|
| }
|
|
|
| // iterator adapters
|
|
|
| -template <typename HashTableType, typename Traits> struct HashTableConstIteratorAdapter {
|
| - STACK_ALLOCATED();
|
| - HashTableConstIteratorAdapter() {}
|
| - HashTableConstIteratorAdapter(const typename HashTableType::const_iterator& impl) : m_impl(impl) {}
|
| - typedef typename Traits::IteratorConstGetType GetType;
|
| - typedef typename HashTableType::ValueTraits::IteratorConstGetType SourceGetType;
|
| -
|
| - GetType get() const { return const_cast<GetType>(SourceGetType(m_impl.get())); }
|
| - typename Traits::IteratorConstReferenceType operator*() const { return Traits::getToReferenceConstConversion(get()); }
|
| - GetType operator->() const { return get(); }
|
| -
|
| - HashTableConstIteratorAdapter& operator++() { ++m_impl; return *this; }
|
| - // postfix ++ intentionally omitted
|
| +template <typename HashTableType, typename Traits>
|
| +struct HashTableConstIteratorAdapter {
|
| + STACK_ALLOCATED();
|
| + HashTableConstIteratorAdapter() {}
|
| + HashTableConstIteratorAdapter(
|
| + const typename HashTableType::const_iterator& impl)
|
| + : m_impl(impl) {}
|
| + typedef typename Traits::IteratorConstGetType GetType;
|
| + typedef
|
| + typename HashTableType::ValueTraits::IteratorConstGetType SourceGetType;
|
| +
|
| + GetType get() const {
|
| + return const_cast<GetType>(SourceGetType(m_impl.get()));
|
| + }
|
| + typename Traits::IteratorConstReferenceType operator*() const {
|
| + return Traits::getToReferenceConstConversion(get());
|
| + }
|
| + GetType operator->() const { return get(); }
|
| +
|
| + HashTableConstIteratorAdapter& operator++() {
|
| + ++m_impl;
|
| + return *this;
|
| + }
|
| + // postfix ++ intentionally omitted
|
|
|
| - typename HashTableType::const_iterator m_impl;
|
| + typename HashTableType::const_iterator m_impl;
|
| };
|
|
|
| -template <typename HashTableType, typename Traits> struct HashTableIteratorAdapter {
|
| - STACK_ALLOCATED();
|
| - typedef typename Traits::IteratorGetType GetType;
|
| - typedef typename HashTableType::ValueTraits::IteratorGetType SourceGetType;
|
| -
|
| - HashTableIteratorAdapter() {}
|
| - HashTableIteratorAdapter(const typename HashTableType::iterator& impl) : m_impl(impl) {}
|
| -
|
| - GetType get() const { return const_cast<GetType>(SourceGetType(m_impl.get())); }
|
| - typename Traits::IteratorReferenceType operator*() const { return Traits::getToReferenceConversion(get()); }
|
| - GetType operator->() const { return get(); }
|
| -
|
| - HashTableIteratorAdapter& operator++() { ++m_impl; return *this; }
|
| - // postfix ++ intentionally omitted
|
| +template <typename HashTableType, typename Traits>
|
| +struct HashTableIteratorAdapter {
|
| + STACK_ALLOCATED();
|
| + typedef typename Traits::IteratorGetType GetType;
|
| + typedef typename HashTableType::ValueTraits::IteratorGetType SourceGetType;
|
| +
|
| + HashTableIteratorAdapter() {}
|
| + HashTableIteratorAdapter(const typename HashTableType::iterator& impl)
|
| + : m_impl(impl) {}
|
| +
|
| + GetType get() const {
|
| + return const_cast<GetType>(SourceGetType(m_impl.get()));
|
| + }
|
| + typename Traits::IteratorReferenceType operator*() const {
|
| + return Traits::getToReferenceConversion(get());
|
| + }
|
| + GetType operator->() const { return get(); }
|
| +
|
| + HashTableIteratorAdapter& operator++() {
|
| + ++m_impl;
|
| + return *this;
|
| + }
|
| + // postfix ++ intentionally omitted
|
|
|
| - operator HashTableConstIteratorAdapter<HashTableType, Traits>()
|
| - {
|
| - typename HashTableType::const_iterator i = m_impl;
|
| - return i;
|
| - }
|
| + operator HashTableConstIteratorAdapter<HashTableType, Traits>() {
|
| + typename HashTableType::const_iterator i = m_impl;
|
| + return i;
|
| + }
|
|
|
| - typename HashTableType::iterator m_impl;
|
| + typename HashTableType::iterator m_impl;
|
| };
|
|
|
| template <typename T, typename U>
|
| -inline bool operator==(const HashTableConstIteratorAdapter<T, U>& a, const HashTableConstIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl == b.m_impl;
|
| +inline bool operator==(const HashTableConstIteratorAdapter<T, U>& a,
|
| + const HashTableConstIteratorAdapter<T, U>& b) {
|
| + return a.m_impl == b.m_impl;
|
| }
|
|
|
| template <typename T, typename U>
|
| -inline bool operator!=(const HashTableConstIteratorAdapter<T, U>& a, const HashTableConstIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl != b.m_impl;
|
| +inline bool operator!=(const HashTableConstIteratorAdapter<T, U>& a,
|
| + const HashTableConstIteratorAdapter<T, U>& b) {
|
| + return a.m_impl != b.m_impl;
|
| }
|
|
|
| template <typename T, typename U>
|
| -inline bool operator==(const HashTableIteratorAdapter<T, U>& a, const HashTableIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl == b.m_impl;
|
| +inline bool operator==(const HashTableIteratorAdapter<T, U>& a,
|
| + const HashTableIteratorAdapter<T, U>& b) {
|
| + return a.m_impl == b.m_impl;
|
| }
|
|
|
| template <typename T, typename U>
|
| -inline bool operator!=(const HashTableIteratorAdapter<T, U>& a, const HashTableIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl != b.m_impl;
|
| +inline bool operator!=(const HashTableIteratorAdapter<T, U>& a,
|
| + const HashTableIteratorAdapter<T, U>& b) {
|
| + return a.m_impl != b.m_impl;
|
| }
|
|
|
| // All 4 combinations of ==, != and Const,non const.
|
| template <typename T, typename U>
|
| -inline bool operator==(const HashTableConstIteratorAdapter<T, U>& a, const HashTableIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl == b.m_impl;
|
| +inline bool operator==(const HashTableConstIteratorAdapter<T, U>& a,
|
| + const HashTableIteratorAdapter<T, U>& b) {
|
| + return a.m_impl == b.m_impl;
|
| }
|
|
|
| template <typename T, typename U>
|
| -inline bool operator!=(const HashTableConstIteratorAdapter<T, U>& a, const HashTableIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl != b.m_impl;
|
| +inline bool operator!=(const HashTableConstIteratorAdapter<T, U>& a,
|
| + const HashTableIteratorAdapter<T, U>& b) {
|
| + return a.m_impl != b.m_impl;
|
| }
|
|
|
| template <typename T, typename U>
|
| -inline bool operator==(const HashTableIteratorAdapter<T, U>& a, const HashTableConstIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl == b.m_impl;
|
| +inline bool operator==(const HashTableIteratorAdapter<T, U>& a,
|
| + const HashTableConstIteratorAdapter<T, U>& b) {
|
| + return a.m_impl == b.m_impl;
|
| }
|
|
|
| template <typename T, typename U>
|
| -inline bool operator!=(const HashTableIteratorAdapter<T, U>& a, const HashTableConstIteratorAdapter<T, U>& b)
|
| -{
|
| - return a.m_impl != b.m_impl;
|
| +inline bool operator!=(const HashTableIteratorAdapter<T, U>& a,
|
| + const HashTableConstIteratorAdapter<T, U>& b) {
|
| + return a.m_impl != b.m_impl;
|
| }
|
|
|
| template <typename Collection1, typename Collection2>
|
| -inline void removeAll(Collection1& collection, const Collection2& toBeRemoved)
|
| -{
|
| - if (collection.isEmpty() || toBeRemoved.isEmpty())
|
| - return;
|
| - typedef typename Collection2::const_iterator CollectionIterator;
|
| - CollectionIterator end(toBeRemoved.end());
|
| - for (CollectionIterator it(toBeRemoved.begin()); it != end; ++it)
|
| - collection.remove(*it);
|
| +inline void removeAll(Collection1& collection, const Collection2& toBeRemoved) {
|
| + if (collection.isEmpty() || toBeRemoved.isEmpty())
|
| + return;
|
| + typedef typename Collection2::const_iterator CollectionIterator;
|
| + CollectionIterator end(toBeRemoved.end());
|
| + for (CollectionIterator it(toBeRemoved.begin()); it != end; ++it)
|
| + collection.remove(*it);
|
| }
|
|
|
| -} // namespace WTF
|
| +} // namespace WTF
|
|
|
| #include "wtf/HashIterators.h"
|
|
|
| -#endif // WTF_HashTable_h
|
| +#endif // WTF_HashTable_h
|
|
|