xref: /freebsd/lib/libthr/thread/thr_umtx.c (revision 07b202a847e08b9cd976e40315cda902ced1c84a)
1 /*
2  * Copyright (c) 2005 David Xu <davidxu@freebsd.org>
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice unmodified, this list of conditions, and the following
10  *    disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25  *
26  * $FreeBSD$
27  *
28  */
29 
30 #include "thr_private.h"
31 #include "thr_umtx.h"
32 
33 #ifndef HAS__UMTX_OP_ERR
34 int _umtx_op_err(void *obj, int op, u_long val, void *uaddr, void *uaddr2)
35 {
36 	if (_umtx_op(obj, op, val, uaddr, uaddr2) == -1)
37 		return (errno);
38 	return (0);
39 }
40 #endif
41 
42 void
43 _thr_umutex_init(struct umutex *mtx)
44 {
45 	static struct umutex default_mtx = DEFAULT_UMUTEX;
46 
47 	*mtx = default_mtx;
48 }
49 
50 void
51 _thr_urwlock_init(struct urwlock *rwl)
52 {
53 	static struct urwlock default_rwl = DEFAULT_URWLOCK;
54 	*rwl = default_rwl;
55 }
56 
57 int
58 __thr_umutex_lock(struct umutex *mtx, uint32_t id)
59 {
60 	uint32_t owner;
61 
62 	if ((mtx->m_flags & (UMUTEX_PRIO_PROTECT | UMUTEX_PRIO_INHERIT)) == 0) {
63 		for (;;) {
64 			/* wait in kernel */
65 			_umtx_op_err(mtx, UMTX_OP_MUTEX_WAIT, 0, 0, 0);
66 
67 			owner = mtx->m_owner;
68 			if ((owner & ~UMUTEX_CONTESTED) == 0 &&
69 			     atomic_cmpset_acq_32(&mtx->m_owner, owner, id|owner))
70 				return (0);
71 		}
72 	}
73 
74 	return	_umtx_op_err(mtx, UMTX_OP_MUTEX_LOCK, 0, 0, 0);
75 }
76 
77 #define SPINLOOPS 1000
78 
79 int
80 __thr_umutex_lock_spin(struct umutex *mtx, uint32_t id)
81 {
82 	uint32_t owner;
83 
84 	if (!_thr_is_smp)
85 		return __thr_umutex_lock(mtx, id);
86 
87 	if ((mtx->m_flags & (UMUTEX_PRIO_PROTECT | UMUTEX_PRIO_INHERIT)) == 0) {
88 		for (;;) {
89 			int count = SPINLOOPS;
90 			while (count--) {
91 				owner = mtx->m_owner;
92 				if ((owner & ~UMUTEX_CONTESTED) == 0) {
93 					if (atomic_cmpset_acq_32(
94 					    &mtx->m_owner,
95 					    owner, id|owner)) {
96 						return (0);
97 					}
98 				}
99 				CPU_SPINWAIT;
100 			}
101 
102 			/* wait in kernel */
103 			_umtx_op_err(mtx, UMTX_OP_MUTEX_WAIT, 0, 0, 0);
104 		}
105 	}
106 
107 	return	_umtx_op_err(mtx, UMTX_OP_MUTEX_LOCK, 0, 0, 0);
108 }
109 
110 int
111 __thr_umutex_timedlock(struct umutex *mtx, uint32_t id,
112 	const struct timespec *abstime)
113 {
114 	struct _umtx_time *tm_p, timeout;
115 	size_t tm_size;
116 	uint32_t owner;
117 	int ret;
118 
119 	if (abstime == NULL) {
120 		tm_p = NULL;
121 		tm_size = 0;
122 	} else {
123 		timeout._clockid = CLOCK_REALTIME;
124 		timeout._flags = UMTX_ABSTIME;
125 		timeout._timeout = *abstime;
126 		tm_p = &timeout;
127 		tm_size = sizeof(timeout);
128 	}
129 
130 	for (;;) {
131 		if ((mtx->m_flags & (UMUTEX_PRIO_PROTECT | UMUTEX_PRIO_INHERIT)) == 0) {
132 
133 			/* wait in kernel */
134 			ret = _umtx_op_err(mtx, UMTX_OP_MUTEX_WAIT, 0,
135 				 (void *)tm_size, __DECONST(void *, tm_p));
136 
137 			/* now try to lock it */
138 			owner = mtx->m_owner;
139 			if ((owner & ~UMUTEX_CONTESTED) == 0 &&
140 			     atomic_cmpset_acq_32(&mtx->m_owner, owner, id|owner))
141 				return (0);
142 		} else {
143 			ret = _umtx_op_err(mtx, UMTX_OP_MUTEX_LOCK, 0,
144 				 (void *)tm_size, __DECONST(void *, tm_p));
145 			if (ret == 0)
146 				break;
147 		}
148 		if (ret == ETIMEDOUT)
149 			break;
150 	}
151 	return (ret);
152 }
153 
154 int
155 __thr_umutex_unlock(struct umutex *mtx, uint32_t id)
156 {
157 #ifndef __ia64__
158 	/* XXX this logic has a race-condition on ia64. */
159 	if ((mtx->m_flags & (UMUTEX_PRIO_PROTECT | UMUTEX_PRIO_INHERIT)) == 0) {
160 		atomic_cmpset_rel_32(&mtx->m_owner, id | UMUTEX_CONTESTED, UMUTEX_CONTESTED);
161 		return _umtx_op_err(mtx, UMTX_OP_MUTEX_WAKE, 0, 0, 0);
162 	}
163 #endif /* __ia64__ */
164 	return _umtx_op_err(mtx, UMTX_OP_MUTEX_UNLOCK, 0, 0, 0);
165 }
166 
167 int
168 __thr_umutex_trylock(struct umutex *mtx)
169 {
170 	return _umtx_op_err(mtx, UMTX_OP_MUTEX_TRYLOCK, 0, 0, 0);
171 }
172 
173 int
174 __thr_umutex_set_ceiling(struct umutex *mtx, uint32_t ceiling,
175 	uint32_t *oldceiling)
176 {
177 	return _umtx_op_err(mtx, UMTX_OP_SET_CEILING, ceiling, oldceiling, 0);
178 }
179 
180 int
181 _thr_umtx_wait(volatile long *mtx, long id, const struct timespec *timeout)
182 {
183 	if (timeout && (timeout->tv_sec < 0 || (timeout->tv_sec == 0 &&
184 		timeout->tv_nsec <= 0)))
185 		return (ETIMEDOUT);
186 	return _umtx_op_err(__DEVOLATILE(void *, mtx), UMTX_OP_WAIT, id, 0,
187 		__DECONST(void*, timeout));
188 }
189 
190 int
191 _thr_umtx_wait_uint(volatile u_int *mtx, u_int id, const struct timespec *timeout, int shared)
192 {
193 	if (timeout && (timeout->tv_sec < 0 || (timeout->tv_sec == 0 &&
194 		timeout->tv_nsec <= 0)))
195 		return (ETIMEDOUT);
196 	return _umtx_op_err(__DEVOLATILE(void *, mtx),
197 			shared ? UMTX_OP_WAIT_UINT : UMTX_OP_WAIT_UINT_PRIVATE, id, 0,
198 			__DECONST(void*, timeout));
199 }
200 
201 int
202 _thr_umtx_timedwait_uint(volatile u_int *mtx, u_int id, int clockid,
203 	const struct timespec *abstime, int shared)
204 {
205 	struct _umtx_time *tm_p, timeout;
206 	size_t tm_size;
207 
208 	if (abstime == NULL) {
209 		tm_p = NULL;
210 		tm_size = 0;
211 	} else {
212 		timeout._clockid = CLOCK_REALTIME;
213 		timeout._flags = UMTX_ABSTIME;
214 		timeout._timeout = *abstime;
215 		tm_p = &timeout;
216 		tm_size = sizeof(timeout);
217 	}
218 
219 	return _umtx_op_err(__DEVOLATILE(void *, mtx),
220 		shared ? UMTX_OP_WAIT_UINT : UMTX_OP_WAIT_UINT_PRIVATE, id,
221 		(void *)tm_size, __DECONST(void *, tm_p));
222 }
223 
224 int
225 _thr_umtx_wake(volatile void *mtx, int nr_wakeup, int shared)
226 {
227 	return _umtx_op_err(__DEVOLATILE(void *, mtx), shared ? UMTX_OP_WAKE : UMTX_OP_WAKE_PRIVATE,
228 		nr_wakeup, 0, 0);
229 }
230 
231 void
232 _thr_ucond_init(struct ucond *cv)
233 {
234 	bzero(cv, sizeof(struct ucond));
235 }
236 
237 int
238 _thr_ucond_wait(struct ucond *cv, struct umutex *m,
239 	const struct timespec *timeout, int flags)
240 {
241 	if (timeout && (timeout->tv_sec < 0 || (timeout->tv_sec == 0 &&
242 	    timeout->tv_nsec <= 0))) {
243 		struct pthread *curthread = _get_curthread();
244 		_thr_umutex_unlock(m, TID(curthread));
245                 return (ETIMEDOUT);
246 	}
247 	return _umtx_op_err(cv, UMTX_OP_CV_WAIT, flags,
248 		     m, __DECONST(void*, timeout));
249 }
250 
251 int
252 _thr_ucond_signal(struct ucond *cv)
253 {
254 	if (!cv->c_has_waiters)
255 		return (0);
256 	return _umtx_op_err(cv, UMTX_OP_CV_SIGNAL, 0, NULL, NULL);
257 }
258 
259 int
260 _thr_ucond_broadcast(struct ucond *cv)
261 {
262 	if (!cv->c_has_waiters)
263 		return (0);
264 	return _umtx_op_err(cv, UMTX_OP_CV_BROADCAST, 0, NULL, NULL);
265 }
266 
267 int
268 __thr_rwlock_rdlock(struct urwlock *rwlock, int flags,
269 	const struct timespec *tsp)
270 {
271 	struct _umtx_time timeout, *tm_p;
272 	size_t tm_size;
273 
274 	if (tsp == NULL) {
275 		tm_p = NULL;
276 		tm_size = 0;
277 	} else {
278 		timeout._timeout = *tsp;
279 		timeout._flags = UMTX_ABSTIME;
280 		timeout._clockid = CLOCK_REALTIME;
281 		tm_p = &timeout;
282 		tm_size = sizeof(timeout);
283 	}
284 	return _umtx_op_err(rwlock, UMTX_OP_RW_RDLOCK, flags, (void *)tm_size, tm_p);
285 }
286 
287 int
288 __thr_rwlock_wrlock(struct urwlock *rwlock, const struct timespec *tsp)
289 {
290 	struct _umtx_time timeout, *tm_p;
291 	size_t tm_size;
292 
293 	if (tsp == NULL) {
294 		tm_p = NULL;
295 		tm_size = 0;
296 	} else {
297 		timeout._timeout = *tsp;
298 		timeout._flags = UMTX_ABSTIME;
299 		timeout._clockid = CLOCK_REALTIME;
300 		tm_p = &timeout;
301 		tm_size = sizeof(timeout);
302 	}
303 	return _umtx_op_err(rwlock, UMTX_OP_RW_WRLOCK, 0, (void *)tm_size, tm_p);
304 }
305 
306 int
307 __thr_rwlock_unlock(struct urwlock *rwlock)
308 {
309 	return _umtx_op_err(rwlock, UMTX_OP_RW_UNLOCK, 0, NULL, NULL);
310 }
311 
312 void
313 _thr_rwl_rdlock(struct urwlock *rwlock)
314 {
315 	int ret;
316 
317 	for (;;) {
318 		if (_thr_rwlock_tryrdlock(rwlock, URWLOCK_PREFER_READER) == 0)
319 			return;
320 		ret = __thr_rwlock_rdlock(rwlock, URWLOCK_PREFER_READER, NULL);
321 		if (ret == 0)
322 			return;
323 		if (ret != EINTR)
324 			PANIC("rdlock error");
325 	}
326 }
327 
328 void
329 _thr_rwl_wrlock(struct urwlock *rwlock)
330 {
331 	int ret;
332 
333 	for (;;) {
334 		if (_thr_rwlock_trywrlock(rwlock) == 0)
335 			return;
336 		ret = __thr_rwlock_wrlock(rwlock, NULL);
337 		if (ret == 0)
338 			return;
339 		if (ret != EINTR)
340 			PANIC("wrlock error");
341 	}
342 }
343 
344 void
345 _thr_rwl_unlock(struct urwlock *rwlock)
346 {
347 	if (_thr_rwlock_unlock(rwlock))
348 		PANIC("unlock error");
349 }
350