OLD | NEW |
---|---|
(Empty) | |
1 /* | |
2 * Copyright (c) 2014 The Native Client Authors. All rights reserved. | |
3 * Use of this source code is governed by a BSD-style license that can be | |
4 * found in the LICENSE file. | |
5 */ | |
6 | |
7 /* | |
8 * Native Client mutex implementation | |
9 */ | |
10 | |
11 #include <errno.h> | |
12 #include <limits.h> | |
13 | |
14 #include "native_client/src/untrusted/nacl/nacl_irt.h" | |
15 #include "native_client/src/untrusted/pthread/pthread.h" | |
16 #include "native_client/src/untrusted/pthread/pthread_internal.h" | |
17 #include "native_client/src/untrusted/pthread/pthread_types.h" | |
18 | |
19 /* | |
20 * This implemtation is based on the one found in the bionic C library | |
Mark Seaborn
2014/10/02 19:58:51
typo
Sam Clegg
2014/10/02 23:12:01
Done.
| |
21 * which is part of the Android Open Source Project | |
22 * (See: libc/bionic/pthread_rwlock.cpp) | |
23 * The following steps were used to port it to NaCl: | |
24 * - convert from C++ to C (mostly replace 'bool') | |
25 * - convert futex calls to __libnacl_irt_futex | |
26 * - remove unsuppored features such as PTHREAD_PROCESS_PRIVATE/SHARED. | |
Mark Seaborn
2014/10/02 19:58:51
typo
Sam Clegg
2014/10/02 23:12:00
Done.
| |
27 */ | |
28 | |
29 /* | |
30 * | |
31 * Technical note: | |
32 * | |
33 * Possible states of a read/write lock: | |
34 * | |
35 * - no readers and no writer (unlocked) | |
36 * - one or more readers sharing the lock at the same time (read-locked) | |
37 * - one writer holding the lock (write-lock) | |
38 * | |
39 * Additionally: | |
40 * - trying to get the write-lock while there are any readers blocks | |
41 * - trying to get the read-lock while there is a writer blocks | |
42 * - a single thread can acquire the lock multiple times in read mode | |
43 * | |
44 * - Posix states that behavior is undefined (may deadlock) if a thread tries | |
45 * to acquire the lock | |
46 * - in write mode while already holding the lock (whether in read or write | |
47 * mode) | |
48 * - in read mode while already holding the lock in write mode. | |
49 * - This implementation will return EDEADLK in "write after write" and "read | |
50 * after write" cases and will deadlock in write after read case. | |
51 */ | |
52 | |
53 int pthread_rwlockattr_init(pthread_rwlockattr_t *attr) { | |
54 *attr = PTHREAD_PROCESS_PRIVATE; | |
55 return 0; | |
56 } | |
57 | |
58 int pthread_rwlockattr_destroy(pthread_rwlockattr_t *attr) { | |
59 *attr = -1; | |
60 return 0; | |
61 } | |
62 | |
63 int pthread_rwlockattr_setpshared(pthread_rwlockattr_t *attr, int pshared) { | |
64 switch (pshared) { | |
65 case PTHREAD_PROCESS_PRIVATE: | |
66 case PTHREAD_PROCESS_SHARED: | |
67 *attr = pshared; | |
68 return 0; | |
69 default: | |
70 return EINVAL; | |
71 } | |
72 } | |
73 | |
74 int pthread_rwlockattr_getpshared(const pthread_rwlockattr_t *attr, | |
75 int *pshared) { | |
76 *pshared = *attr; | |
77 return 0; | |
78 } | |
79 | |
80 int pthread_rwlock_init(pthread_rwlock_t *rwlock, | |
81 const pthread_rwlockattr_t *attr) { | |
82 if (attr != NULL) { | |
83 switch (*attr) { | |
84 case PTHREAD_PROCESS_SHARED: | |
85 case PTHREAD_PROCESS_PRIVATE: | |
86 rwlock->attr= *attr; | |
87 break; | |
88 default: | |
89 return EINVAL; | |
90 } | |
91 } | |
92 | |
93 rwlock->state = 0; | |
94 rwlock->pending_readers = 0; | |
95 rwlock->pending_writers = 0; | |
96 rwlock->writer_thread_id = 0; | |
97 | |
98 return 0; | |
99 } | |
100 | |
101 int pthread_rwlock_destroy(pthread_rwlock_t* rwlock) { | |
102 if (rwlock->state != 0) { | |
103 return EBUSY; | |
104 } | |
105 return 0; | |
106 } | |
107 | |
108 static int __pthread_rwlock_timedrdlock(pthread_rwlock_t *rwlock, | |
109 const struct timespec *abs_timeout) { | |
110 if (__predict_false(pthread_self() == rwlock->writer_thread_id)) { | |
111 return EDEADLK; | |
112 } | |
113 | |
114 int done = 0; | |
115 do { | |
116 /* | |
117 * This is actually a race read as there's nothing that guarantees the | |
Mark Seaborn
2014/10/02 19:58:51
This is not true with PNaCl, where "volatile" make
Sam Clegg
2014/10/02 23:12:00
Acknowledged.
| |
118 * atomicity of integer reads / writes. However, in practice this "never" | |
119 * happens so until we switch to C++11 this should work fine. The same | |
120 * applies in the other places this idiom is used. | |
121 */ | |
122 int32_t cur_state = rwlock->state; /* C++11 relaxed atomic read */ | |
Mark Seaborn
2014/10/02 19:58:51
This isn't in C++. :-)
Sam Clegg
2014/10/02 23:12:00
Acknowledged.
| |
123 if (__predict_true(cur_state >= 0)) { | |
Mark Seaborn
2014/10/02 19:58:51
I don't know if nacl-gcc has __predict. Please ru
Sam Clegg
2014/10/02 23:12:00
Acknowledged.
| |
124 /* Add as an extra reader. */ | |
125 done = __sync_bool_compare_and_swap(&rwlock->state, | |
126 cur_state, | |
127 cur_state + 1); | |
128 } else { | |
129 /* | |
130 * Owner holds it in write mode, hang up. | |
131 * To avoid losing wake ups the pending_readers update and the state read | |
132 * should be sequentially consistent. (currently enforced by | |
133 * __sync_fetch_and_add which creates a full barrier) | |
134 */ | |
135 __sync_fetch_and_add(&rwlock->pending_readers, 1); | |
136 int ret = __libnacl_irt_futex.futex_wait_abs(&rwlock->state, | |
137 cur_state, | |
138 abs_timeout); | |
139 __sync_fetch_and_sub(&rwlock->pending_readers, 1); | |
140 if (ret == -ETIMEDOUT) { | |
Mark Seaborn
2014/10/02 19:58:50
"ret == ETIMEDOUT". Please add test coverage.
Sam Clegg
2014/10/02 23:12:00
Done.
Sam Clegg
2014/10/02 23:12:00
Done.
| |
141 return ETIMEDOUT; | |
142 } | |
143 } | |
144 } while (!done); | |
145 | |
146 return 0; | |
147 } | |
148 | |
149 static int __pthread_rwlock_timedwrlock(pthread_rwlock_t *rwlock, | |
150 const struct timespec *abs_timeout) { | |
151 pthread_t self = pthread_self(); | |
152 if (__predict_false(self == rwlock->writer_thread_id)) { | |
153 return EDEADLK; | |
154 } | |
155 | |
156 int done = 0; | |
157 do { | |
158 int32_t cur_state = rwlock->state; | |
159 if (__predict_true(cur_state == 0)) { | |
160 /* Change state from 0 to -1. */ | |
161 done = __sync_bool_compare_and_swap(&rwlock->state, | |
162 0 /* cur state */, | |
163 -1 /* new state */); | |
164 } else { | |
165 /* | |
166 * Failed to acquire, hang up. | |
167 * To avoid losing wake ups the pending_writers update and the state read | |
168 * should be sequentially consistent. (currently enforced by | |
Mark Seaborn
2014/10/02 19:58:51
Capitalise as "Currently" and add "."
Sam Clegg
2014/10/02 23:12:00
Done.
| |
169 * __sync_fetch_and_add which creates a full barrier) | |
170 */ | |
171 __sync_fetch_and_add(&rwlock->pending_writers, 1); | |
172 int ret = __libnacl_irt_futex.futex_wait_abs(&rwlock->state, | |
173 cur_state, | |
174 abs_timeout); | |
175 __sync_fetch_and_sub(&rwlock->pending_writers, 1); | |
176 if (ret == -ETIMEDOUT) { | |
177 return ETIMEDOUT; | |
178 } | |
179 } | |
180 } while (!done); | |
181 | |
182 rwlock->writer_thread_id = self; | |
183 return 0; | |
184 } | |
185 | |
186 int pthread_rwlock_rdlock(pthread_rwlock_t *rwlock) { | |
187 return __pthread_rwlock_timedrdlock(rwlock, NULL); | |
188 } | |
189 | |
190 int pthread_rwlock_timedrdlock(pthread_rwlock_t *rwlock, | |
191 const struct timespec *abs_timeout) { | |
192 return __pthread_rwlock_timedrdlock(rwlock, abs_timeout); | |
193 } | |
194 | |
195 int pthread_rwlock_tryrdlock(pthread_rwlock_t *rwlock) { | |
196 int32_t cur_state = rwlock->state; | |
197 if ((cur_state >= 0) && | |
198 __sync_bool_compare_and_swap(&rwlock->state, cur_state, cur_state + 1)) { | |
199 return 0; | |
200 } | |
201 return EBUSY; | |
202 } | |
203 | |
204 int pthread_rwlock_wrlock(pthread_rwlock_t *rwlock) { | |
205 return __pthread_rwlock_timedwrlock(rwlock, NULL); | |
206 } | |
207 | |
208 int pthread_rwlock_timedwrlock(pthread_rwlock_t *rwlock, | |
209 const struct timespec *abs_timeout) { | |
210 return __pthread_rwlock_timedwrlock(rwlock, abs_timeout); | |
Mark Seaborn
2014/10/02 19:58:51
pthread_rwlock_timedwrlock() and __pthread_rwlock_
Sam Clegg
2014/10/02 23:12:00
Done.
| |
211 } | |
212 | |
213 int pthread_rwlock_trywrlock(pthread_rwlock_t *rwlock) { | |
214 int32_t cur_state = rwlock->state; | |
215 if ((cur_state == 0) && | |
216 __sync_bool_compare_and_swap(&rwlock->state, | |
217 0 /* cur state */, | |
218 -1 /* new state */)) { | |
219 rwlock->writer_thread_id = pthread_self(); | |
220 return 0; | |
221 } | |
222 return EBUSY; | |
223 } | |
224 | |
225 int pthread_rwlock_unlock(pthread_rwlock_t *rwlock) { | |
226 pthread_t self = pthread_self(); | |
227 int done = 0; | |
228 do { | |
229 int32_t cur_state = rwlock->state; | |
230 if (cur_state == 0) { | |
231 return EPERM; | |
232 } | |
233 if (cur_state == -1) { | |
234 if (rwlock->writer_thread_id != self) { | |
235 return EPERM; | |
236 } | |
237 /* We're no longer the owner. */ | |
238 rwlock->writer_thread_id = 0; | |
239 /* | |
240 * Change state from -1 to 0. | |
241 * We use __sync_bool_compare_and_swap to achieve sequential consistency | |
242 * of the state store and the following pendingX loads. A simple store | |
243 * with memory_order_release semantics is not enough to guarantee that the | |
244 * pendingX loads are not reordered before the store (which may lead to a | |
245 * lost wakeup). | |
246 */ | |
247 __sync_bool_compare_and_swap(&rwlock->state, | |
248 -1 /* cur state*/, | |
249 0 /* new state */); | |
250 | |
251 /* Wake any waiters. */ | |
252 if (__predict_false(rwlock->pending_readers > 0 | |
253 || rwlock->pending_writers > 0)) { | |
254 __libnacl_irt_futex.futex_wake(&rwlock->state, INT_MAX, NULL); | |
255 } | |
256 done = 1; | |
257 } else { /* cur_state > 0 */ | |
258 /* | |
259 * Reduce state by 1. | |
260 * See the comment above on why we need __sync_bool_compare_and_swap. | |
261 */ | |
262 done = __sync_bool_compare_and_swap(&rwlock->state, | |
263 cur_state, | |
264 cur_state - 1); | |
265 if (done && (cur_state - 1) == 0) { | |
266 /* There are no more readers, wake any waiters. */ | |
267 if (__predict_false(rwlock->pending_readers > 0 | |
Mark Seaborn
2014/10/02 19:58:51
I can't see any distinction between pending_reader
Sam Clegg
2014/10/02 23:12:00
Thats what Hans Boehm said too :)
https://android-
| |
268 || rwlock->pending_writers > 0)) { | |
269 __libnacl_irt_futex.futex_wake(&rwlock->state, INT_MAX, NULL); | |
Mark Seaborn
2014/10/02 19:58:51
Waking all the waiters doesn't seem very efficient
Sam Clegg
2014/10/02 23:12:00
The change is here:
https://android-review.googles
| |
270 } | |
271 } | |
272 } | |
273 } while (!done); | |
274 | |
275 return 0; | |
276 } | |
OLD | NEW |