| Index: include/ports/SkAtomics_sync.h
|
| diff --git a/include/ports/SkAtomics_sync.h b/include/ports/SkAtomics_sync.h
|
| deleted file mode 100644
|
| index 02b1e580725d9f949ded0b3c6861f07de16fa8dc..0000000000000000000000000000000000000000
|
| --- a/include/ports/SkAtomics_sync.h
|
| +++ /dev/null
|
| @@ -1,73 +0,0 @@
|
| -/*
|
| - * Copyright 2015 Google Inc.
|
| - *
|
| - * Use of this source code is governed by a BSD-style license that can be
|
| - * found in the LICENSE file.
|
| - */
|
| -
|
| -#ifndef SkAtomics_sync_DEFINED
|
| -#define SkAtomics_sync_DEFINED
|
| -
|
| -// This file is mostly a shim. We'd like to delete it. Please don't put much
|
| -// effort into maintaining it, and if you find bugs in it, the right fix is to
|
| -// delete this file and upgrade your compiler to something that supports
|
| -// __atomic builtins or std::atomic.
|
| -
|
| -static inline void barrier(sk_memory_order mo) {
|
| - asm volatile("" : : : "memory"); // Prevents the compiler from reordering code.
|
| - #if SK_CPU_X86
|
| - // On x86, we generally don't need an extra memory barrier for loads or stores.
|
| - if (sk_memory_order_seq_cst == mo) { __sync_synchronize(); }
|
| - #else
|
| - // On other platforms (e.g. ARM) we do unless the memory order is relaxed.
|
| - if (sk_memory_order_relaxed != mo) { __sync_synchronize(); }
|
| - #endif
|
| -}
|
| -
|
| -// These barriers only support our majority use cases: acquire and relaxed loads, release stores.
|
| -// For anything more complicated, please consider deleting this file and upgrading your compiler.
|
| -
|
| -template <typename T>
|
| -T sk_atomic_load(const T* ptr, sk_memory_order mo) {
|
| - T val = *ptr;
|
| - barrier(mo);
|
| - return val;
|
| -}
|
| -
|
| -template <typename T>
|
| -void sk_atomic_store(T* ptr, T val, sk_memory_order mo) {
|
| - barrier(mo);
|
| - *ptr = val;
|
| -}
|
| -
|
| -template <typename T>
|
| -T sk_atomic_fetch_add(T* ptr, T val, sk_memory_order) {
|
| - return __sync_fetch_and_add(ptr, val);
|
| -}
|
| -
|
| -template <typename T>
|
| -T sk_atomic_fetch_sub(T* ptr, T val, sk_memory_order) {
|
| - return __sync_fetch_and_sub(ptr, val);
|
| -}
|
| -
|
| -template <typename T>
|
| -bool sk_atomic_compare_exchange(T* ptr, T* expected, T desired, sk_memory_order, sk_memory_order) {
|
| - T prev = __sync_val_compare_and_swap(ptr, *expected, desired);
|
| - if (prev == *expected) {
|
| - return true;
|
| - }
|
| - *expected = prev;
|
| - return false;
|
| -}
|
| -
|
| -template <typename T>
|
| -T sk_atomic_exchange(T* ptr, T val, sk_memory_order) {
|
| - // There is no __sync exchange. Emulate it with a CAS loop.
|
| - T prev;
|
| - do {
|
| - prev = sk_atomic_load(ptr);
|
| - } while(!sk_atomic_compare_exchange(ptr, &prev, val));
|
| - return prev;
|
| -}
|
| -
|
| -#endif//SkAtomics_sync_DEFINED
|
|
|