Index: third_party/WebKit/Source/wtf/Atomics.h |
diff --git a/third_party/WebKit/Source/wtf/Atomics.h b/third_party/WebKit/Source/wtf/Atomics.h |
index b08cdcd205b8e0a635261a3803d4916a5517ff17..2fb3d19daa839db73b54eee9b9ff8eaf148de545 100644 |
--- a/third_party/WebKit/Source/wtf/Atomics.h |
+++ b/third_party/WebKit/Source/wtf/Atomics.h |
@@ -1,402 +1,9 @@ |
-/* |
- * Copyright (C) 2007, 2008, 2010, 2012 Apple Inc. All rights reserved. |
- * Copyright (C) 2007 Justin Haygood (jhaygood@reaktix.com) |
- * |
- * Redistribution and use in source and binary forms, with or without |
- * modification, are permitted provided that the following conditions |
- * are met: |
- * |
- * 1. Redistributions of source code must retain the above copyright |
- * notice, this list of conditions and the following disclaimer. |
- * 2. Redistributions in binary form must reproduce the above copyright |
- * notice, this list of conditions and the following disclaimer in the |
- * documentation and/or other materials provided with the distribution. |
- * 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of |
- * its contributors may be used to endorse or promote products derived |
- * from this software without specific prior written permission. |
- * |
- * THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY |
- * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED |
- * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
- * DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY |
- * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES |
- * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; |
- * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND |
- * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
- * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
- * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
- */ |
+// Copyright 2017 The Chromium Authors. All rights reserved. |
+// Use of this source code is governed by a BSD-style license that can be |
+// found in the LICENSE file. |
-#ifndef Atomics_h |
-#define Atomics_h |
+#include "platform/wtf/Atomics.h" |
-#include "wtf/AddressSanitizer.h" |
-#include "wtf/Assertions.h" |
-#include "wtf/CPU.h" |
- |
-#include <stdint.h> |
- |
-#if COMPILER(MSVC) |
-#include <windows.h> |
-#endif |
- |
-#if defined(THREAD_SANITIZER) |
-#include <sanitizer/tsan_interface_atomic.h> |
-#endif |
- |
-#if defined(ADDRESS_SANITIZER) |
-#include <sanitizer/asan_interface.h> |
-#endif |
- |
-namespace WTF { |
- |
-#if COMPILER(MSVC) |
- |
-// atomicAdd returns the result of the addition. |
-ALWAYS_INLINE int atomicAdd(int volatile* addend, int increment) { |
- return InterlockedExchangeAdd(reinterpret_cast<long volatile*>(addend), |
- static_cast<long>(increment)) + |
- increment; |
-} |
-ALWAYS_INLINE unsigned atomicAdd(unsigned volatile* addend, |
- unsigned increment) { |
- return InterlockedExchangeAdd(reinterpret_cast<long volatile*>(addend), |
- static_cast<long>(increment)) + |
- increment; |
-} |
-#if defined(_WIN64) |
-ALWAYS_INLINE unsigned long long atomicAdd(unsigned long long volatile* addend, |
- unsigned long long increment) { |
- return InterlockedExchangeAdd64(reinterpret_cast<long long volatile*>(addend), |
- static_cast<long long>(increment)) + |
- increment; |
-} |
-#endif |
- |
-// atomicSubtract returns the result of the subtraction. |
-ALWAYS_INLINE int atomicSubtract(int volatile* addend, int decrement) { |
- return InterlockedExchangeAdd(reinterpret_cast<long volatile*>(addend), |
- static_cast<long>(-decrement)) - |
- decrement; |
-} |
-ALWAYS_INLINE unsigned atomicSubtract(unsigned volatile* addend, |
- unsigned decrement) { |
- return InterlockedExchangeAdd(reinterpret_cast<long volatile*>(addend), |
- -static_cast<long>(decrement)) - |
- decrement; |
-} |
-#if defined(_WIN64) |
-ALWAYS_INLINE unsigned long long atomicSubtract( |
- unsigned long long volatile* addend, |
- unsigned long long decrement) { |
- return InterlockedExchangeAdd64(reinterpret_cast<long long volatile*>(addend), |
- -static_cast<long long>(decrement)) - |
- decrement; |
-} |
-#endif |
- |
-ALWAYS_INLINE int atomicIncrement(int volatile* addend) { |
- return InterlockedIncrement(reinterpret_cast<long volatile*>(addend)); |
-} |
-ALWAYS_INLINE int atomicDecrement(int volatile* addend) { |
- return InterlockedDecrement(reinterpret_cast<long volatile*>(addend)); |
-} |
- |
-ALWAYS_INLINE int64_t atomicIncrement(int64_t volatile* addend) { |
- return InterlockedIncrement64(reinterpret_cast<long long volatile*>(addend)); |
-} |
-ALWAYS_INLINE int64_t atomicDecrement(int64_t volatile* addend) { |
- return InterlockedDecrement64(reinterpret_cast<long long volatile*>(addend)); |
-} |
- |
-ALWAYS_INLINE int atomicTestAndSetToOne(int volatile* ptr) { |
- int ret = InterlockedExchange(reinterpret_cast<long volatile*>(ptr), 1); |
- DCHECK(!ret || ret == 1); |
- return ret; |
-} |
- |
-ALWAYS_INLINE void atomicSetOneToZero(int volatile* ptr) { |
- DCHECK_EQ(*ptr, 1); |
- InterlockedExchange(reinterpret_cast<long volatile*>(ptr), 0); |
-} |
- |
-#else |
- |
-// atomicAdd returns the result of the addition. |
-ALWAYS_INLINE int atomicAdd(int volatile* addend, int increment) { |
- return __sync_add_and_fetch(addend, increment); |
-} |
-ALWAYS_INLINE unsigned atomicAdd(unsigned volatile* addend, |
- unsigned increment) { |
- return __sync_add_and_fetch(addend, increment); |
-} |
-ALWAYS_INLINE unsigned long atomicAdd(unsigned long volatile* addend, |
- unsigned long increment) { |
- return __sync_add_and_fetch(addend, increment); |
-} |
-// atomicSubtract returns the result of the subtraction. |
-ALWAYS_INLINE int atomicSubtract(int volatile* addend, int decrement) { |
- return __sync_sub_and_fetch(addend, decrement); |
-} |
-ALWAYS_INLINE unsigned atomicSubtract(unsigned volatile* addend, |
- unsigned decrement) { |
- return __sync_sub_and_fetch(addend, decrement); |
-} |
-ALWAYS_INLINE unsigned long atomicSubtract(unsigned long volatile* addend, |
- unsigned long decrement) { |
- return __sync_sub_and_fetch(addend, decrement); |
-} |
- |
-ALWAYS_INLINE int atomicIncrement(int volatile* addend) { |
- return atomicAdd(addend, 1); |
-} |
-ALWAYS_INLINE int atomicDecrement(int volatile* addend) { |
- return atomicSubtract(addend, 1); |
-} |
- |
-ALWAYS_INLINE int64_t atomicIncrement(int64_t volatile* addend) { |
- return __sync_add_and_fetch(addend, 1); |
-} |
-ALWAYS_INLINE int64_t atomicDecrement(int64_t volatile* addend) { |
- return __sync_sub_and_fetch(addend, 1); |
-} |
- |
-ALWAYS_INLINE int atomicTestAndSetToOne(int volatile* ptr) { |
- int ret = __sync_lock_test_and_set(ptr, 1); |
- DCHECK(!ret || ret == 1); |
- return ret; |
-} |
- |
-ALWAYS_INLINE void atomicSetOneToZero(int volatile* ptr) { |
- DCHECK_EQ(*ptr, 1); |
- __sync_lock_release(ptr); |
-} |
-#endif |
- |
-#if defined(THREAD_SANITIZER) |
-// The definitions below assume an LP64 data model. This is fine because |
-// TSan is only supported on x86_64 Linux. |
-#if CPU(64BIT) && OS(LINUX) |
-ALWAYS_INLINE void releaseStore(volatile int* ptr, int value) { |
- __tsan_atomic32_store(ptr, value, __tsan_memory_order_release); |
-} |
-ALWAYS_INLINE void releaseStore(volatile unsigned* ptr, unsigned value) { |
- __tsan_atomic32_store(reinterpret_cast<volatile int*>(ptr), |
- static_cast<int>(value), __tsan_memory_order_release); |
-} |
-ALWAYS_INLINE void releaseStore(volatile long* ptr, long value) { |
- __tsan_atomic64_store(reinterpret_cast<volatile __tsan_atomic64*>(ptr), |
- static_cast<__tsan_atomic64>(value), |
- __tsan_memory_order_release); |
-} |
-ALWAYS_INLINE void releaseStore(volatile unsigned long* ptr, |
- unsigned long value) { |
- __tsan_atomic64_store(reinterpret_cast<volatile __tsan_atomic64*>(ptr), |
- static_cast<__tsan_atomic64>(value), |
- __tsan_memory_order_release); |
-} |
-ALWAYS_INLINE void releaseStore(volatile unsigned long long* ptr, |
- unsigned long long value) { |
- __tsan_atomic64_store(reinterpret_cast<volatile __tsan_atomic64*>(ptr), |
- static_cast<__tsan_atomic64>(value), |
- __tsan_memory_order_release); |
-} |
-ALWAYS_INLINE void releaseStore(void* volatile* ptr, void* value) { |
- __tsan_atomic64_store(reinterpret_cast<volatile __tsan_atomic64*>(ptr), |
- reinterpret_cast<__tsan_atomic64>(value), |
- __tsan_memory_order_release); |
-} |
-ALWAYS_INLINE int acquireLoad(volatile const int* ptr) { |
- return __tsan_atomic32_load(ptr, __tsan_memory_order_acquire); |
-} |
-ALWAYS_INLINE unsigned acquireLoad(volatile const unsigned* ptr) { |
- return static_cast<unsigned>(__tsan_atomic32_load( |
- reinterpret_cast<volatile const int*>(ptr), __tsan_memory_order_acquire)); |
-} |
-ALWAYS_INLINE long acquireLoad(volatile const long* ptr) { |
- return static_cast<long>(__tsan_atomic64_load( |
- reinterpret_cast<volatile const __tsan_atomic64*>(ptr), |
- __tsan_memory_order_acquire)); |
-} |
-ALWAYS_INLINE unsigned long acquireLoad(volatile const unsigned long* ptr) { |
- return static_cast<unsigned long>(__tsan_atomic64_load( |
- reinterpret_cast<volatile const __tsan_atomic64*>(ptr), |
- __tsan_memory_order_acquire)); |
-} |
-ALWAYS_INLINE void* acquireLoad(void* volatile const* ptr) { |
- return reinterpret_cast<void*>(__tsan_atomic64_load( |
- reinterpret_cast<volatile const __tsan_atomic64*>(ptr), |
- __tsan_memory_order_acquire)); |
-} |
- |
-// Do not use noBarrierStore/noBarrierLoad for synchronization. |
-ALWAYS_INLINE void noBarrierStore(volatile float* ptr, float value) { |
- static_assert(sizeof(int) == sizeof(float), |
- "int and float are different sizes"); |
- union { |
- int ivalue; |
- float fvalue; |
- } u; |
- u.fvalue = value; |
- __tsan_atomic32_store(reinterpret_cast<volatile __tsan_atomic32*>(ptr), |
- u.ivalue, __tsan_memory_order_relaxed); |
-} |
- |
-ALWAYS_INLINE float noBarrierLoad(volatile const float* ptr) { |
- static_assert(sizeof(int) == sizeof(float), |
- "int and float are different sizes"); |
- union { |
- int ivalue; |
- float fvalue; |
- } u; |
- u.ivalue = __tsan_atomic32_load(reinterpret_cast<volatile const int*>(ptr), |
- __tsan_memory_order_relaxed); |
- return u.fvalue; |
-} |
-#endif |
- |
-#else // defined(THREAD_SANITIZER) |
- |
-#if CPU(X86) || CPU(X86_64) |
-// Only compiler barrier is needed. |
-#if COMPILER(MSVC) |
-// Starting from Visual Studio 2005 compiler guarantees acquire and release |
-// semantics for operations on volatile variables. See MSDN entry for |
-// MemoryBarrier macro. |
-#define MEMORY_BARRIER() |
-#else |
-#define MEMORY_BARRIER() __asm__ __volatile__("" : : : "memory") |
-#endif |
-#else |
-// Fallback to the compiler intrinsic on all other platforms. |
-#define MEMORY_BARRIER() __sync_synchronize() |
-#endif |
- |
-ALWAYS_INLINE void releaseStore(volatile int* ptr, int value) { |
- MEMORY_BARRIER(); |
- *ptr = value; |
-} |
-ALWAYS_INLINE void releaseStore(volatile unsigned* ptr, unsigned value) { |
- MEMORY_BARRIER(); |
- *ptr = value; |
-} |
-ALWAYS_INLINE void releaseStore(volatile long* ptr, long value) { |
- MEMORY_BARRIER(); |
- *ptr = value; |
-} |
-ALWAYS_INLINE void releaseStore(volatile unsigned long* ptr, |
- unsigned long value) { |
- MEMORY_BARRIER(); |
- *ptr = value; |
-} |
-#if CPU(64BIT) |
-ALWAYS_INLINE void releaseStore(volatile unsigned long long* ptr, |
- unsigned long long value) { |
- MEMORY_BARRIER(); |
- *ptr = value; |
-} |
-#endif |
-ALWAYS_INLINE void releaseStore(void* volatile* ptr, void* value) { |
- MEMORY_BARRIER(); |
- *ptr = value; |
-} |
- |
-ALWAYS_INLINE int acquireLoad(volatile const int* ptr) { |
- int value = *ptr; |
- MEMORY_BARRIER(); |
- return value; |
-} |
-ALWAYS_INLINE unsigned acquireLoad(volatile const unsigned* ptr) { |
- unsigned value = *ptr; |
- MEMORY_BARRIER(); |
- return value; |
-} |
-ALWAYS_INLINE long acquireLoad(volatile const long* ptr) { |
- long value = *ptr; |
- MEMORY_BARRIER(); |
- return value; |
-} |
-ALWAYS_INLINE unsigned long acquireLoad(volatile const unsigned long* ptr) { |
- unsigned long value = *ptr; |
- MEMORY_BARRIER(); |
- return value; |
-} |
-#if CPU(64BIT) |
-ALWAYS_INLINE unsigned long long acquireLoad( |
- volatile const unsigned long long* ptr) { |
- unsigned long long value = *ptr; |
- MEMORY_BARRIER(); |
- return value; |
-} |
-#endif |
-ALWAYS_INLINE void* acquireLoad(void* volatile const* ptr) { |
- void* value = *ptr; |
- MEMORY_BARRIER(); |
- return value; |
-} |
- |
-// Do not use noBarrierStore/noBarrierLoad for synchronization. |
-ALWAYS_INLINE void noBarrierStore(volatile float* ptr, float value) { |
- *ptr = value; |
-} |
- |
-ALWAYS_INLINE float noBarrierLoad(volatile const float* ptr) { |
- float value = *ptr; |
- return value; |
-} |
- |
-#if defined(ADDRESS_SANITIZER) |
- |
-NO_SANITIZE_ADDRESS ALWAYS_INLINE void asanUnsafeReleaseStore( |
- volatile unsigned* ptr, |
- unsigned value) { |
- MEMORY_BARRIER(); |
- *ptr = value; |
-} |
- |
-NO_SANITIZE_ADDRESS ALWAYS_INLINE unsigned asanUnsafeAcquireLoad( |
- volatile const unsigned* ptr) { |
- unsigned value = *ptr; |
- MEMORY_BARRIER(); |
- return value; |
-} |
- |
-#endif // defined(ADDRESS_SANITIZER) |
- |
-#undef MEMORY_BARRIER |
- |
-#endif |
- |
-#if !defined(ADDRESS_SANITIZER) |
- |
-ALWAYS_INLINE void asanUnsafeReleaseStore(volatile unsigned* ptr, |
- unsigned value) { |
- releaseStore(ptr, value); |
-} |
- |
-ALWAYS_INLINE unsigned asanUnsafeAcquireLoad(volatile const unsigned* ptr) { |
- return acquireLoad(ptr); |
-} |
- |
-#endif |
- |
-} // namespace WTF |
- |
-using WTF::atomicAdd; |
-using WTF::atomicSubtract; |
-using WTF::atomicDecrement; |
-using WTF::atomicIncrement; |
-using WTF::atomicTestAndSetToOne; |
-using WTF::atomicSetOneToZero; |
-using WTF::acquireLoad; |
-using WTF::releaseStore; |
-using WTF::noBarrierLoad; |
-using WTF::noBarrierStore; |
- |
-// These methods allow loading from and storing to poisoned memory. Only |
-// use these methods if you know what you are doing since they will |
-// silence use-after-poison errors from ASan. |
-using WTF::asanUnsafeAcquireLoad; |
-using WTF::asanUnsafeReleaseStore; |
- |
-#endif // Atomics_h |
+// The contents of this header was moved to platform/wtf as part of |
+// WTF migration project. See the following post for details: |
+// https://groups.google.com/a/chromium.org/d/msg/blink-dev/tLdAZCTlcAA/bYXVT8gYCAAJ |