Lines Matching defs:waiter

163  * @wqh:	The wake queue head from which to get the next lock waiter
276 * @waiter: the pre-initialized rt_mutex_waiter
280 * Starts the rt_mutex acquire; it enqueues the @waiter and does deadlock
283 * NOTE: does _NOT_ remove the @waiter on failure; must either call
294 struct rt_mutex_waiter *waiter,
306 ret = task_blocks_on_rt_mutex(lock, waiter, task, NULL,
314 * pi chain. Let the waiter sort it out.
325 * @waiter: the pre-initialized rt_mutex_waiter
328 * Starts the rt_mutex acquire; it enqueues the @waiter and does deadlock
331 * NOTE: unlike __rt_mutex_start_proxy_lock this _DOES_ remove the @waiter
342 struct rt_mutex_waiter *waiter,
349 ret = __rt_mutex_start_proxy_lock(lock, waiter, task, &wake_q);
351 remove_waiter(lock, waiter);
365 * @waiter: the pre-initialized rt_mutex_waiter
379 struct rt_mutex_waiter *waiter)
386 ret = rt_mutex_slowlock_block(lock, NULL, TASK_INTERRUPTIBLE, to, waiter, NULL);
388 * try_to_take_rt_mutex() sets the waiter bit unconditionally. We might
400 * @waiter: the pre-initialized rt_mutex_waiter
418 struct rt_mutex_waiter *waiter)
430 * we will own the lock and it will have removed the waiter. If we
434 try_to_take_rt_mutex(lock, current, waiter);
440 remove_waiter(lock, waiter);
444 * try_to_take_rt_mutex() sets the waiter bit unconditionally. We might
461 struct rt_mutex_waiter *waiter;
467 waiter = task->pi_blocked_on;
468 if (!waiter || rt_waiter_node_equal(&waiter->tree, task_to_waiter_node(task))) {
472 next_lock = waiter->lock;
483 * Performs the wakeup of the top-waiter and re-enables preemption.