Index: src/base/atomicops.h |
diff --git a/src/base/atomicops.h b/src/base/atomicops.h |
index 927ebbee11d6cd5b05b812d8589e049fcab2209d..7fb07006d7ec19d1463a936b7aafa789a616f729 100644 |
--- a/src/base/atomicops.h |
+++ b/src/base/atomicops.h |
@@ -14,10 +14,10 @@ |
// do not know what you are doing, avoid these routines, and use a Mutex. |
// |
// It is incorrect to make direct assignments to/from an atomic variable. |
-// You should use one of the Load or Store routines. The NoBarrier |
-// versions are provided when no barriers are needed: |
-// NoBarrier_Store() |
-// NoBarrier_Load() |
+// You should use one of the Load or Store routines. The Relaxed versions |
+// are provided when no fences are needed: |
+// Relaxed_Store() |
+// Relaxed_Load() |
// Although there are currently no compiler enforcement, you are encouraged |
// to use these. |
// |
@@ -74,17 +74,16 @@ typedef intptr_t AtomicWord; |
// Always return the old value of "*ptr" |
// |
// This routine implies no memory barriers. |
-Atomic32 NoBarrier_CompareAndSwap(volatile Atomic32* ptr, |
- Atomic32 old_value, |
- Atomic32 new_value); |
+Atomic32 Relaxed_CompareAndSwap(volatile Atomic32* ptr, Atomic32 old_value, |
+ Atomic32 new_value); |
// Atomically store new_value into *ptr, returning the previous value held in |
// *ptr. This routine implies no memory barriers. |
-Atomic32 NoBarrier_AtomicExchange(volatile Atomic32* ptr, Atomic32 new_value); |
+Atomic32 Relaxed_AtomicExchange(volatile Atomic32* ptr, Atomic32 new_value); |
// Atomically increment *ptr by "increment". Returns the new value of |
// *ptr with the increment applied. This routine implies no memory barriers. |
-Atomic32 NoBarrier_AtomicIncrement(volatile Atomic32* ptr, Atomic32 increment); |
+Atomic32 Relaxed_AtomicIncrement(volatile Atomic32* ptr, Atomic32 increment); |
Atomic32 Barrier_AtomicIncrement(volatile Atomic32* ptr, |
Atomic32 increment); |
@@ -106,21 +105,20 @@ Atomic32 Release_CompareAndSwap(volatile Atomic32* ptr, |
Atomic32 new_value); |
void MemoryBarrier(); |
-void NoBarrier_Store(volatile Atomic8* ptr, Atomic8 value); |
-void NoBarrier_Store(volatile Atomic32* ptr, Atomic32 value); |
+void Relaxed_Store(volatile Atomic8* ptr, Atomic8 value); |
+void Relaxed_Store(volatile Atomic32* ptr, Atomic32 value); |
void Release_Store(volatile Atomic32* ptr, Atomic32 value); |
-Atomic8 NoBarrier_Load(volatile const Atomic8* ptr); |
-Atomic32 NoBarrier_Load(volatile const Atomic32* ptr); |
+Atomic8 Relaxed_Load(volatile const Atomic8* ptr); |
+Atomic32 Relaxed_Load(volatile const Atomic32* ptr); |
Atomic32 Acquire_Load(volatile const Atomic32* ptr); |
// 64-bit atomic operations (only available on 64-bit processors). |
#ifdef V8_HOST_ARCH_64_BIT |
-Atomic64 NoBarrier_CompareAndSwap(volatile Atomic64* ptr, |
- Atomic64 old_value, |
- Atomic64 new_value); |
-Atomic64 NoBarrier_AtomicExchange(volatile Atomic64* ptr, Atomic64 new_value); |
-Atomic64 NoBarrier_AtomicIncrement(volatile Atomic64* ptr, Atomic64 increment); |
+Atomic64 Relaxed_CompareAndSwap(volatile Atomic64* ptr, Atomic64 old_value, |
+ Atomic64 new_value); |
+Atomic64 Relaxed_AtomicExchange(volatile Atomic64* ptr, Atomic64 new_value); |
+Atomic64 Relaxed_AtomicIncrement(volatile Atomic64* ptr, Atomic64 increment); |
Atomic64 Barrier_AtomicIncrement(volatile Atomic64* ptr, Atomic64 increment); |
Atomic64 Acquire_CompareAndSwap(volatile Atomic64* ptr, |
@@ -129,9 +127,9 @@ Atomic64 Acquire_CompareAndSwap(volatile Atomic64* ptr, |
Atomic64 Release_CompareAndSwap(volatile Atomic64* ptr, |
Atomic64 old_value, |
Atomic64 new_value); |
-void NoBarrier_Store(volatile Atomic64* ptr, Atomic64 value); |
+void Relaxed_Store(volatile Atomic64* ptr, Atomic64 value); |
void Release_Store(volatile Atomic64* ptr, Atomic64 value); |
-Atomic64 NoBarrier_Load(volatile const Atomic64* ptr); |
+Atomic64 Relaxed_Load(volatile const Atomic64* ptr); |
Atomic64 Acquire_Load(volatile const Atomic64* ptr); |
#endif // V8_HOST_ARCH_64_BIT |