| /linux/kernel/locking/ |
| H A D | ww_mutex.h | 313 struct ww_acquire_ctx *ww_ctx, struct wake_q_head *wake_q) in __ww_mutex_die() argument 328 wake_q_add(wake_q, waiter->task); in __ww_mutex_die() 344 struct wake_q_head *wake_q) in __ww_mutex_wound() argument 387 wake_q_add(wake_q, owner); in __ww_mutex_wound() 409 struct wake_q_head *wake_q) in __ww_mutex_check_waiters() argument 422 if (__ww_mutex_die(lock, cur, ww_ctx, wake_q) || in __ww_mutex_check_waiters() 423 __ww_mutex_wound(lock, cur->ww_ctx, ww_ctx, wake_q)) in __ww_mutex_check_waiters() 435 DEFINE_WAKE_Q(wake_q); in ww_mutex_set_context_fastpath() 468 __ww_mutex_check_waiters(&lock->base, ctx, &wake_q); in ww_mutex_set_context_fastpath() 471 wake_up_q(&wake_q); in ww_mutex_set_context_fastpath() [all …]
|
| H A D | rtmutex.c | 39 struct wake_q_head *wake_q) in __ww_mutex_add_waiter() argument 46 struct wake_q_head *wake_q) in __ww_mutex_check_waiters() argument 1214 struct wake_q_head *wake_q) in task_blocks_on_rt_mutex() argument 1260 res = __ww_mutex_add_waiter(waiter, rtm, ww_ctx, wake_q); in task_blocks_on_rt_mutex() 1304 raw_spin_unlock_irq_wake(&lock->wait_lock, wake_q); in task_blocks_on_rt_mutex() 1622 struct wake_q_head *wake_q) in rt_mutex_slowlock_block() argument 1658 raw_spin_unlock_irq_wake(&lock->wait_lock, wake_q); in rt_mutex_slowlock_block() 1712 struct wake_q_head *wake_q) in __rt_mutex_slowlock() argument 1726 __ww_mutex_check_waiters(rtm, ww_ctx, wake_q); in __rt_mutex_slowlock() 1737 ret = task_blocks_on_rt_mutex(lock, waiter, current, ww_ctx, chwalk, wake_q); in __rt_mutex_slowlock() [all …]
|
| H A D | mutex.c | 614 DEFINE_WAKE_Q(wake_q); in __mutex_lock_common() 666 __ww_mutex_check_waiters(lock, ww_ctx, &wake_q); in __mutex_lock_common() 686 ret = __ww_mutex_add_waiter(&waiter, lock, ww_ctx, &wake_q); in __mutex_lock_common() 724 raw_spin_unlock_irqrestore_wake(&lock->wait_lock, flags, &wake_q); in __mutex_lock_common() 782 __ww_mutex_check_waiters(lock, ww_ctx, &wake_q); in __mutex_lock_common() 797 raw_spin_unlock_irqrestore_wake(&lock->wait_lock, flags, &wake_q); in __mutex_lock_common() 808 raw_spin_unlock_irqrestore_wake(&lock->wait_lock, flags, &wake_q); in __mutex_lock_common() 985 DEFINE_WAKE_Q(wake_q); in __mutex_unlock_slowpath() 1023 wake_q_add(&wake_q, next); in __mutex_unlock_slowpath() 1029 raw_spin_unlock_irqrestore_wake(&lock->wait_lock, flags, &wake_q); in __mutex_unlock_slowpath()
|
| /linux/ipc/ |
| H A D | sem.c | 787 struct wake_q_head *wake_q) in wake_up_sem_queue_prepare() argument 796 wake_q_add_safe(wake_q, sleeper); in wake_up_sem_queue_prepare() 855 struct wake_q_head *wake_q) in wake_const_ops() argument 874 wake_up_sem_queue_prepare(q, error, wake_q); in wake_const_ops() 894 int nsops, struct wake_q_head *wake_q) in do_smart_wakeup_zero() argument 907 semop_completed |= wake_const_ops(sma, num, wake_q); in do_smart_wakeup_zero() 918 semop_completed |= wake_const_ops(sma, i, wake_q); in do_smart_wakeup_zero() 927 semop_completed |= wake_const_ops(sma, -1, wake_q); in do_smart_wakeup_zero() 949 static int update_queue(struct sem_array *sma, int semnum, struct wake_q_head *wake_q) in update_queue() argument 986 do_smart_wakeup_zero(sma, q->sops, q->nsops, wake_q); in update_queue() [all …]
|
| H A D | msg.c | 213 struct wake_q_head *wake_q, bool kill) in ss_wakeup() argument 245 wake_q_add(wake_q, mss->tsk); in ss_wakeup() 250 struct wake_q_head *wake_q) in expunge_all() argument 261 wake_q_add_safe(wake_q, r_tsk); in expunge_all() 279 DEFINE_WAKE_Q(wake_q); in freeque() 281 expunge_all(msq, -EIDRM, &wake_q); in freeque() 282 ss_wakeup(msq, &wake_q, true); in freeque() 285 wake_up_q(&wake_q); in freeque() 432 DEFINE_WAKE_Q(wake_q); in msgctl_down() 452 expunge_all(msq, -EAGAIN, &wake_q); in msgctl_down() [all …]
|
| H A D | mqueue.c | 993 static inline void __pipelined_op(struct wake_q_head *wake_q, in __pipelined_op() argument 1004 wake_q_add_safe(wake_q, task); in __pipelined_op() 1010 static inline void pipelined_send(struct wake_q_head *wake_q, in pipelined_send() argument 1016 __pipelined_op(wake_q, info, receiver); in pipelined_send() 1021 static inline void pipelined_receive(struct wake_q_head *wake_q, in pipelined_receive() argument 1034 __pipelined_op(wake_q, info, sender); in pipelined_receive() 1049 DEFINE_WAKE_Q(wake_q); in do_mq_timedsend() 1123 pipelined_send(&wake_q, info, msg_ptr, receiver); in do_mq_timedsend() 1135 wake_up_q(&wake_q); in do_mq_timedsend() 1209 DEFINE_WAKE_Q(wake_q); in do_mq_timedreceive() [all …]
|
| /linux/kernel/futex/ |
| H A D | requeue.c | 390 DEFINE_WAKE_Q(wake_q); in futex_requeue() 613 this->wake(&wake_q, this); in futex_requeue() 695 wake_up_q(&wake_q); in futex_requeue()
|
| H A D | pi.c | 929 DEFINE_WAKE_Q(wake_q); in futex_lock_pi() 1041 ret = __rt_mutex_start_proxy_lock(&q.pi_state->pi_mutex, &rt_waiter, current, &wake_q); in futex_lock_pi() 1042 raw_spin_unlock_irq_wake(&q.pi_state->pi_mutex.wait_lock, &wake_q); in futex_lock_pi()
|
| /linux/io_uring/ |
| H A D | futex.c | 157 static void io_futex_wakev_fn(struct wake_q_head *wake_q, struct futex_q *q) in io_futex_wakev_fn() argument 210 static void io_futex_wake_fn(struct wake_q_head *wake_q, struct futex_q *q) in io_futex_wake_fn() argument
|
| /linux/include/linux/ |
| H A D | sched.h | 1230 struct wake_q_node wake_q; member
|
| /linux/kernel/sched/ |
| H A D | core.c | 1078 struct wake_q_node *node = &task->wake_q; in __wake_q_add() 1148 task = container_of(node, struct task_struct, wake_q); in wake_up_q() 1151 WRITE_ONCE(task->wake_q.next, NULL); in wake_up_q()
|
| /linux/kernel/ |
| H A D | fork.c | 974 tsk->wake_q.next = NULL; in dup_task_struct()
|