| Index: include/core/SkAtomics.h
|
| diff --git a/include/core/SkAtomics.h b/include/core/SkAtomics.h
|
| deleted file mode 100644
|
| index e947d1a9a47775d1673761ab9f29657cb2e757a0..0000000000000000000000000000000000000000
|
| --- a/include/core/SkAtomics.h
|
| +++ /dev/null
|
| @@ -1,136 +0,0 @@
|
| -/*
|
| - * Copyright 2015 Google Inc.
|
| - *
|
| - * Use of this source code is governed by a BSD-style license that can be
|
| - * found in the LICENSE file.
|
| - */
|
| -
|
| -#ifndef SkAtomics_DEFINED
|
| -#define SkAtomics_DEFINED
|
| -
|
| -// This file is not part of the public Skia API.
|
| -#include "SkTypes.h"
|
| -
|
| -enum sk_memory_order {
|
| - sk_memory_order_relaxed,
|
| - sk_memory_order_consume,
|
| - sk_memory_order_acquire,
|
| - sk_memory_order_release,
|
| - sk_memory_order_acq_rel,
|
| - sk_memory_order_seq_cst,
|
| -};
|
| -
|
| -template <typename T>
|
| -T sk_atomic_load(const T*, sk_memory_order = sk_memory_order_seq_cst);
|
| -
|
| -template <typename T>
|
| -void sk_atomic_store(T*, T, sk_memory_order = sk_memory_order_seq_cst);
|
| -
|
| -template <typename T>
|
| -T sk_atomic_fetch_add(T*, T, sk_memory_order = sk_memory_order_seq_cst);
|
| -
|
| -template <typename T>
|
| -T sk_atomic_fetch_sub(T*, T, sk_memory_order = sk_memory_order_seq_cst);
|
| -
|
| -template <typename T>
|
| -bool sk_atomic_compare_exchange(T*, T* expected, T desired,
|
| - sk_memory_order success = sk_memory_order_seq_cst,
|
| - sk_memory_order failure = sk_memory_order_seq_cst);
|
| -
|
| -template <typename T>
|
| -T sk_atomic_exchange(T*, T, sk_memory_order = sk_memory_order_seq_cst);
|
| -
|
| -// A little wrapper class for small T (think, builtins: int, float, void*) to
|
| -// ensure they're always used atomically. This is our stand-in for std::atomic<T>.
|
| -template <typename T>
|
| -class SkAtomic : SkNoncopyable {
|
| -public:
|
| - SkAtomic() {}
|
| - explicit SkAtomic(const T& val) : fVal(val) {}
|
| -
|
| - // It is essential we return by value rather than by const&. fVal may change at any time.
|
| - T load(sk_memory_order mo = sk_memory_order_seq_cst) const {
|
| - return sk_atomic_load(&fVal, mo);
|
| - }
|
| -
|
| - void store(const T& val, sk_memory_order mo = sk_memory_order_seq_cst) {
|
| - sk_atomic_store(&fVal, val, mo);
|
| - }
|
| -
|
| - // Alias for .load(sk_memory_order_seq_cst).
|
| - operator T() const {
|
| - return this->load();
|
| - }
|
| -
|
| - // Alias for .store(v, sk_memory_order_seq_cst).
|
| - T operator=(const T& v) {
|
| - this->store(v);
|
| - return v;
|
| - }
|
| -
|
| - T fetch_add(const T& val, sk_memory_order mo = sk_memory_order_seq_cst) {
|
| - return sk_atomic_fetch_add(&fVal, val, mo);
|
| - }
|
| -
|
| - T fetch_sub(const T& val, sk_memory_order mo = sk_memory_order_seq_cst) {
|
| - return sk_atomic_fetch_sub(&fVal, val, mo);
|
| - }
|
| -
|
| - bool compare_exchange(T* expected, const T& desired,
|
| - sk_memory_order success = sk_memory_order_seq_cst,
|
| - sk_memory_order failure = sk_memory_order_seq_cst) {
|
| - return sk_atomic_compare_exchange(&fVal, expected, desired, success, failure);
|
| - }
|
| -private:
|
| - T fVal;
|
| -};
|
| -
|
| -// IWYU pragma: begin_exports
|
| -#if defined(_MSC_VER)
|
| - #include "../ports/SkAtomics_std.h"
|
| -#elif !defined(SK_BUILD_FOR_IOS) && defined(__ATOMIC_RELAXED)
|
| - #include "../ports/SkAtomics_atomic.h"
|
| -#else
|
| - #include "../ports/SkAtomics_sync.h"
|
| -#endif
|
| -// IWYU pragma: end_exports
|
| -
|
| -// From here down we have shims for our old atomics API, to be weaned off of.
|
| -// We use the default sequentially-consistent memory order to make things simple
|
| -// and to match the practical reality of our old _sync and _win implementations.
|
| -
|
| -inline int32_t sk_atomic_inc(int32_t* ptr) { return sk_atomic_fetch_add(ptr, +1); }
|
| -inline int32_t sk_atomic_dec(int32_t* ptr) { return sk_atomic_fetch_add(ptr, -1); }
|
| -inline int32_t sk_atomic_add(int32_t* ptr, int32_t v) { return sk_atomic_fetch_add(ptr, v); }
|
| -
|
| -inline int64_t sk_atomic_inc(int64_t* ptr) { return sk_atomic_fetch_add<int64_t>(ptr, +1); }
|
| -
|
| -inline bool sk_atomic_cas(int32_t* ptr, int32_t expected, int32_t desired) {
|
| - return sk_atomic_compare_exchange(ptr, &expected, desired);
|
| -}
|
| -
|
| -inline void* sk_atomic_cas(void** ptr, void* expected, void* desired) {
|
| - (void)sk_atomic_compare_exchange(ptr, &expected, desired);
|
| - return expected;
|
| -}
|
| -
|
| -inline int32_t sk_atomic_conditional_inc(int32_t* ptr) {
|
| - int32_t prev = sk_atomic_load(ptr);
|
| - do {
|
| - if (0 == prev) {
|
| - break;
|
| - }
|
| - } while(!sk_atomic_compare_exchange(ptr, &prev, prev+1));
|
| - return prev;
|
| -}
|
| -
|
| -template <typename T>
|
| -T sk_acquire_load(T* ptr) { return sk_atomic_load(ptr, sk_memory_order_acquire); }
|
| -
|
| -template <typename T>
|
| -void sk_release_store(T* ptr, T val) { sk_atomic_store(ptr, val, sk_memory_order_release); }
|
| -
|
| -inline void sk_membar_acquire__after_atomic_dec() {}
|
| -inline void sk_membar_acquire__after_atomic_conditional_inc() {}
|
| -
|
| -#endif//SkAtomics_DEFINED
|
|
|