OLD | NEW |
1 // Copyright 2013 Red Hat Inc. All rights reserved. | 1 // Copyright 2013 Red Hat Inc. All rights reserved. |
2 // | 2 // |
3 // Redistribution and use in source and binary forms, with or without | 3 // Redistribution and use in source and binary forms, with or without |
4 // modification, are permitted provided that the following conditions are | 4 // modification, are permitted provided that the following conditions are |
5 // met: | 5 // met: |
6 // | 6 // |
7 // * Redistributions of source code must retain the above copyright | 7 // * Redistributions of source code must retain the above copyright |
8 // notice, this list of conditions and the following disclaimer. | 8 // notice, this list of conditions and the following disclaimer. |
9 // * Redistributions in binary form must reproduce the above | 9 // * Redistributions in binary form must reproduce the above |
10 // copyright notice, this list of conditions and the following disclaimer | 10 // copyright notice, this list of conditions and the following disclaimer |
(...skipping 20 matching lines...) Expand all Loading... |
31 #ifndef GOOGLE_PROTOBUF_ATOMICOPS_INTERNALS_GENERIC_GCC_H_ | 31 #ifndef GOOGLE_PROTOBUF_ATOMICOPS_INTERNALS_GENERIC_GCC_H_ |
32 #define GOOGLE_PROTOBUF_ATOMICOPS_INTERNALS_GENERIC_GCC_H_ | 32 #define GOOGLE_PROTOBUF_ATOMICOPS_INTERNALS_GENERIC_GCC_H_ |
33 | 33 |
34 namespace google { | 34 namespace google { |
35 namespace protobuf { | 35 namespace protobuf { |
36 namespace internal { | 36 namespace internal { |
37 | 37 |
38 inline Atomic32 NoBarrier_CompareAndSwap(volatile Atomic32* ptr, | 38 inline Atomic32 NoBarrier_CompareAndSwap(volatile Atomic32* ptr, |
39 Atomic32 old_value, | 39 Atomic32 old_value, |
40 Atomic32 new_value) { | 40 Atomic32 new_value) { |
41 __atomic_compare_exchange_n(ptr, &old_value, new_value, false, | 41 __atomic_compare_exchange_n(ptr, &old_value, new_value, true, |
42 __ATOMIC_RELAXED, __ATOMIC_RELAXED); | 42 __ATOMIC_RELAXED, __ATOMIC_RELAXED); |
43 return old_value; | 43 return old_value; |
44 } | 44 } |
45 | 45 |
46 inline Atomic32 NoBarrier_AtomicExchange(volatile Atomic32* ptr, | 46 inline Atomic32 NoBarrier_AtomicExchange(volatile Atomic32* ptr, |
47 Atomic32 new_value) { | 47 Atomic32 new_value) { |
48 return __atomic_exchange_n(ptr, new_value, __ATOMIC_RELAXED); | 48 return __atomic_exchange_n(ptr, new_value, __ATOMIC_RELAXED); |
49 } | 49 } |
50 | 50 |
51 inline Atomic32 NoBarrier_AtomicIncrement(volatile Atomic32* ptr, | 51 inline Atomic32 NoBarrier_AtomicIncrement(volatile Atomic32* ptr, |
52 Atomic32 increment) { | 52 Atomic32 increment) { |
53 return __atomic_add_fetch(ptr, increment, __ATOMIC_RELAXED); | 53 return __atomic_add_fetch(ptr, increment, __ATOMIC_RELAXED); |
54 } | 54 } |
55 | 55 |
56 inline Atomic32 Barrier_AtomicIncrement(volatile Atomic32* ptr, | 56 inline Atomic32 Barrier_AtomicIncrement(volatile Atomic32* ptr, |
57 Atomic32 increment) { | 57 Atomic32 increment) { |
58 return __atomic_add_fetch(ptr, increment, __ATOMIC_SEQ_CST); | 58 return __atomic_add_fetch(ptr, increment, __ATOMIC_SEQ_CST); |
59 } | 59 } |
60 | 60 |
61 inline Atomic32 Acquire_CompareAndSwap(volatile Atomic32* ptr, | 61 inline Atomic32 Acquire_CompareAndSwap(volatile Atomic32* ptr, |
62 Atomic32 old_value, | 62 Atomic32 old_value, |
63 Atomic32 new_value) { | 63 Atomic32 new_value) { |
64 __atomic_compare_exchange_n(ptr, &old_value, new_value, false, | 64 __atomic_compare_exchange_n(ptr, &old_value, new_value, true, |
65 __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE); | 65 __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE); |
66 return old_value; | 66 return old_value; |
67 } | 67 } |
68 | 68 |
69 inline Atomic32 Release_CompareAndSwap(volatile Atomic32* ptr, | 69 inline Atomic32 Release_CompareAndSwap(volatile Atomic32* ptr, |
70 Atomic32 old_value, | 70 Atomic32 old_value, |
71 Atomic32 new_value) { | 71 Atomic32 new_value) { |
72 __atomic_compare_exchange_n(ptr, &old_value, new_value, false, | 72 __atomic_compare_exchange_n(ptr, &old_value, new_value, true, |
73 __ATOMIC_RELEASE, __ATOMIC_ACQUIRE); | 73 __ATOMIC_RELEASE, __ATOMIC_ACQUIRE); |
74 return old_value; | 74 return old_value; |
75 } | 75 } |
76 | 76 |
77 inline void NoBarrier_Store(volatile Atomic32* ptr, Atomic32 value) { | 77 inline void NoBarrier_Store(volatile Atomic32* ptr, Atomic32 value) { |
78 __atomic_store_n(ptr, value, __ATOMIC_RELAXED); | 78 __atomic_store_n(ptr, value, __ATOMIC_RELAXED); |
79 } | 79 } |
80 | 80 |
81 inline void MemoryBarrierInternal() { | 81 inline void MemoryBarrier() { |
82 __sync_synchronize(); | 82 __sync_synchronize(); |
83 } | 83 } |
84 | 84 |
85 inline void Acquire_Store(volatile Atomic32* ptr, Atomic32 value) { | 85 inline void Acquire_Store(volatile Atomic32* ptr, Atomic32 value) { |
86 __atomic_store_n(ptr, value, __ATOMIC_SEQ_CST); | 86 __atomic_store_n(ptr, value, __ATOMIC_SEQ_CST); |
87 } | 87 } |
88 | 88 |
89 inline void Release_Store(volatile Atomic32* ptr, Atomic32 value) { | 89 inline void Release_Store(volatile Atomic32* ptr, Atomic32 value) { |
90 __atomic_store_n(ptr, value, __ATOMIC_RELEASE); | 90 __atomic_store_n(ptr, value, __ATOMIC_RELEASE); |
91 } | 91 } |
(...skipping 16 matching lines...) Expand all Loading... |
108 __atomic_store_n(ptr, value, __ATOMIC_RELEASE); | 108 __atomic_store_n(ptr, value, __ATOMIC_RELEASE); |
109 } | 109 } |
110 | 110 |
111 inline Atomic64 Acquire_Load(volatile const Atomic64* ptr) { | 111 inline Atomic64 Acquire_Load(volatile const Atomic64* ptr) { |
112 return __atomic_load_n(ptr, __ATOMIC_ACQUIRE); | 112 return __atomic_load_n(ptr, __ATOMIC_ACQUIRE); |
113 } | 113 } |
114 | 114 |
115 inline Atomic64 Acquire_CompareAndSwap(volatile Atomic64* ptr, | 115 inline Atomic64 Acquire_CompareAndSwap(volatile Atomic64* ptr, |
116 Atomic64 old_value, | 116 Atomic64 old_value, |
117 Atomic64 new_value) { | 117 Atomic64 new_value) { |
118 __atomic_compare_exchange_n(ptr, &old_value, new_value, false, | 118 __atomic_compare_exchange_n(ptr, &old_value, new_value, true, |
119 __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE); | 119 __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE); |
120 return old_value; | 120 return old_value; |
121 } | 121 } |
122 | 122 |
123 inline Atomic64 NoBarrier_CompareAndSwap(volatile Atomic64* ptr, | 123 inline Atomic64 NoBarrier_CompareAndSwap(volatile Atomic64* ptr, |
124 Atomic64 old_value, | 124 Atomic64 old_value, |
125 Atomic64 new_value) { | 125 Atomic64 new_value) { |
126 __atomic_compare_exchange_n(ptr, &old_value, new_value, false, | 126 __atomic_compare_exchange_n(ptr, &old_value, new_value, true, |
127 __ATOMIC_RELAXED, __ATOMIC_RELAXED); | 127 __ATOMIC_RELAXED, __ATOMIC_RELAXED); |
128 return old_value; | 128 return old_value; |
129 } | 129 } |
130 | 130 |
131 inline Atomic64 NoBarrier_AtomicIncrement(volatile Atomic64* ptr, | |
132 Atomic64 increment) { | |
133 return __atomic_add_fetch(ptr, increment, __ATOMIC_RELAXED); | |
134 } | |
135 | |
136 inline void NoBarrier_Store(volatile Atomic64* ptr, Atomic64 value) { | |
137 __atomic_store_n(ptr, value, __ATOMIC_RELAXED); | |
138 } | |
139 | |
140 inline Atomic64 NoBarrier_AtomicExchange(volatile Atomic64* ptr, | |
141 Atomic64 new_value) { | |
142 return __atomic_exchange_n(ptr, new_value, __ATOMIC_RELAXED); | |
143 } | |
144 | |
145 inline Atomic64 NoBarrier_Load(volatile const Atomic64* ptr) { | |
146 return __atomic_load_n(ptr, __ATOMIC_RELAXED); | |
147 } | |
148 | |
149 #endif // defined(__LP64__) | 131 #endif // defined(__LP64__) |
150 | 132 |
151 } // namespace internal | 133 } // namespace internal |
152 } // namespace protobuf | 134 } // namespace protobuf |
153 } // namespace google | 135 } // namespace google |
154 | 136 |
155 #endif // GOOGLE_PROTOBUF_ATOMICOPS_INTERNALS_GENERIC_GCC_H_ | 137 #endif // GOOGLE_PROTOBUF_ATOMICOPS_INTERNALS_GENERIC_GCC_H_ |
OLD | NEW |