Home
last modified time | relevance | path

Searched full:wait (Results 1 – 25 of 3860) sorted by relevance

12345678910>>...155

/linux/drivers/infiniband/hw/hfi1/
H A Diowait.c12 void iowait_set_flag(struct iowait *wait, u32 flag) in iowait_set_flag() argument
14 trace_hfi1_iowait_set(wait, flag); in iowait_set_flag()
15 set_bit(flag, &wait->flags); in iowait_set_flag()
18 bool iowait_flag_set(struct iowait *wait, u32 flag) in iowait_flag_set() argument
20 return test_bit(flag, &wait->flags); in iowait_flag_set()
23 inline void iowait_clear_flag(struct iowait *wait, u32 flag) in iowait_clear_flag() argument
25 trace_hfi1_iowait_clear(wait, flag); in iowait_clear_flag()
26 clear_bit(flag, &wait->flags); in iowait_clear_flag()
30 * iowait_init() - initialize wait structure
31 * @wait: wait struct to initialize
[all …]
H A Dtrace_iowait.h17 TP_PROTO(struct iowait *wait, u32 flag),
18 TP_ARGS(wait, flag),
26 __entry->addr = (unsigned long)wait;
27 __entry->flags = wait->flags;
29 __entry->qpn = iowait_to_qp(wait)->ibqp.qp_num;
41 TP_PROTO(struct iowait *wait, u32 flag),
42 TP_ARGS(wait, flag));
45 TP_PROTO(struct iowait *wait, u32 flag),
46 TP_ARGS(wait, flag));
/linux/drivers/gpu/drm/nouveau/nvif/
H A Dtimer.c26 nvif_timer_wait_test(struct nvif_timer_wait *wait) in nvif_timer_wait_test() argument
28 u64 time = nvif_device_time(wait->device); in nvif_timer_wait_test()
30 if (wait->reads == 0) { in nvif_timer_wait_test()
31 wait->time0 = time; in nvif_timer_wait_test()
32 wait->time1 = time; in nvif_timer_wait_test()
35 if (wait->time1 == time) { in nvif_timer_wait_test()
36 if (WARN_ON(wait->reads++ == 16)) in nvif_timer_wait_test()
39 wait->time1 = time; in nvif_timer_wait_test()
40 wait->reads = 1; in nvif_timer_wait_test()
43 if (wait->time1 - wait->time0 > wait->limit) in nvif_timer_wait_test()
[all …]
/linux/arch/mips/kernel/
H A Didle.c3 * MIPS idle loop and WAIT instruction support.
23 * Not all of the MIPS CPUs have the "wait" instruction available. Moreover,
24 * the implementation of the "wait" feature differs between CPU families. This
25 * points to the function that implements CPU specific wait.
26 * The wait instruction stops the pipeline and reduces the power consumption of
51 " wait \n" in r4k_wait_irqoff()
57 * have any pending stores when the WAIT instruction is executed.
69 " wait \n" in rm7k_wait_irqoff()
75 * Au1 'wait' is only useful when the 32kHz counter is used as timer,
90 " wait \n" in au1k_wait()
[all …]
/linux/fs/
H A Dfs_pin.c16 spin_lock_irq(&pin->wait.lock); in pin_remove()
18 wake_up_locked(&pin->wait); in pin_remove()
19 spin_unlock_irq(&pin->wait.lock); in pin_remove()
32 wait_queue_entry_t wait; in pin_kill() local
38 init_wait(&wait); in pin_kill()
39 spin_lock_irq(&p->wait.lock); in pin_kill()
42 spin_unlock_irq(&p->wait.lock); in pin_kill()
48 spin_unlock_irq(&p->wait.lock); in pin_kill()
52 __add_wait_queue(&p->wait, &wait); in pin_kill()
55 spin_unlock_irq(&p->wait.lock); in pin_kill()
[all …]
/linux/kernel/sched/
H A Dswait.c3 * <linux/swait.h> (simple wait queues ) implementation:
85 void __prepare_to_swait(struct swait_queue_head *q, struct swait_queue *wait) in __prepare_to_swait() argument
87 wait->task = current; in __prepare_to_swait()
88 if (list_empty(&wait->task_list)) in __prepare_to_swait()
89 list_add_tail(&wait->task_list, &q->task_list); in __prepare_to_swait()
92 void prepare_to_swait_exclusive(struct swait_queue_head *q, struct swait_queue *wait, int state) in prepare_to_swait_exclusive() argument
97 __prepare_to_swait(q, wait); in prepare_to_swait_exclusive()
103 long prepare_to_swait_event(struct swait_queue_head *q, struct swait_queue *wait, int state) in prepare_to_swait_event() argument
114 list_del_init(&wait->task_list); in prepare_to_swait_event()
117 __prepare_to_swait(q, wait); in prepare_to_swait_event()
[all …]
H A Dcompletion.c4 * Generic wait-for-completion handler;
25 raw_spin_lock_irqsave(&x->wait.lock, flags); in complete_with_flags()
29 swake_up_locked(&x->wait, wake_flags); in complete_with_flags()
30 raw_spin_unlock_irqrestore(&x->wait.lock, flags); in complete_with_flags()
78 raw_spin_lock_irqsave(&x->wait.lock, flags); in complete_all()
80 swake_up_all_locked(&x->wait); in complete_all()
81 raw_spin_unlock_irqrestore(&x->wait.lock, flags); in complete_all()
90 DECLARE_SWAITQUEUE(wait); in do_wait_for_common()
97 __prepare_to_swait(&x->wait, &wait); in do_wait_for_common()
99 raw_spin_unlock_irq(&x->wait.lock); in do_wait_for_common()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/timer/
H A Dbase.c27 nvkm_timer_wait_test(struct nvkm_timer_wait *wait) in nvkm_timer_wait_test() argument
29 struct nvkm_subdev *subdev = &wait->tmr->subdev; in nvkm_timer_wait_test()
30 u64 time = nvkm_timer_read(wait->tmr); in nvkm_timer_wait_test()
32 if (wait->reads == 0) { in nvkm_timer_wait_test()
33 wait->time0 = time; in nvkm_timer_wait_test()
34 wait->time1 = time; in nvkm_timer_wait_test()
37 if (wait->time1 == time) { in nvkm_timer_wait_test()
38 if (wait->reads++ == 16) { in nvkm_timer_wait_test()
43 wait->time1 = time; in nvkm_timer_wait_test()
44 wait->reads = 1; in nvkm_timer_wait_test()
[all …]
/linux/drivers/gpu/drm/omapdrm/
H A Domap_irq.c23 struct omap_irq_wait *wait; in omap_irq_update() local
28 list_for_each_entry(wait, &priv->wait_list, node) in omap_irq_update()
29 irqmask |= wait->irqmask; in omap_irq_update()
36 static void omap_irq_wait_handler(struct omap_irq_wait *wait) in omap_irq_wait_handler() argument
38 wait->count--; in omap_irq_wait_handler()
39 wake_up(&wait->wq); in omap_irq_wait_handler()
46 struct omap_irq_wait *wait = kzalloc_obj(*wait); in omap_irq_wait_init() local
49 init_waitqueue_head(&wait->wq); in omap_irq_wait_init()
50 wait->irqmask = irqmask; in omap_irq_wait_init()
51 wait->count = count; in omap_irq_wait_init()
[all …]
/linux/drivers/scsi/fnic/
H A Dvnic_dev.c62 int wait);
249 static int vnic_dev_cmd1(struct vnic_dev *vdev, enum vnic_devcmd_cmd cmd, int wait) in vnic_dev_cmd1() argument
283 for (delay = 0; delay < wait; delay++) { in vnic_dev_cmd1()
312 int wait) in vnic_dev_cmd2() argument
379 for (delay = 0; delay < wait; delay++) { in vnic_dev_cmd2()
501 enum vnic_devcmd_cmd cmd, u64 *a0, u64 *a1, int wait) in vnic_dev_cmd_no_proxy() argument
508 err = (*vdev->devcmd_rtn)(vdev, cmd, wait); in vnic_dev_cmd_no_proxy()
518 u64 *a0, u64 *a1, int wait) in vnic_dev_cmd() argument
525 return vnic_dev_cmd_no_proxy(vdev, cmd, a0, a1, wait); in vnic_dev_cmd()
534 int wait = 1000; in vnic_dev_fw_info() local
[all …]
/linux/net/core/
H A Dstream.c21 #include <linux/wait.h>
45 wake_up_interruptible_poll(&wq->wait, EPOLLOUT | in sk_stream_write_space()
54 * sk_stream_wait_connect - Wait for a socket to get into the connected state
55 * @sk: sock to wait on
56 * @timeo_p: for how long to wait
62 DEFINE_WAIT_FUNC(wait, woken_wake_function); in sk_stream_wait_connect()
77 add_wait_queue(sk_sleep(sk), &wait); in sk_stream_wait_connect()
82 ~(TCPF_ESTABLISHED | TCPF_CLOSE_WAIT)), &wait); in sk_stream_wait_connect()
83 remove_wait_queue(sk_sleep(sk), &wait); in sk_stream_wait_connect()
103 DEFINE_WAIT_FUNC(wait, woken_wake_function); in sk_stream_wait_close()
[all …]
/linux/kernel/
H A Dumh.c118 if (!(sub_info->wait & UMH_WAIT_PROC)) in call_usermodehelper_exec_async()
162 if (sub_info->wait & UMH_WAIT_PROC) { in call_usermodehelper_exec_work()
192 * Wait queue head used by usermodehelper_disable() to wait for all running
198 * Used by usermodehelper_read_lock_wait() to wait for usermodehelper_disabled
204 * Time to wait for running_helpers to become zero before the setting of
211 DEFINE_WAIT(wait); in usermodehelper_read_trylock()
216 prepare_to_wait(&usermodehelper_disabled_waitq, &wait, in usermodehelper_read_trylock()
234 finish_wait(&usermodehelper_disabled_waitq, &wait); in usermodehelper_read_trylock()
241 DEFINE_WAIT(wait); in usermodehelper_read_lock_wait()
248 prepare_to_wait(&usermodehelper_disabled_waitq, &wait, in usermodehelper_read_lock_wait()
[all …]
/linux/include/linux/
H A Dswait.h8 #include <linux/wait.h>
12 * Simple waitqueues are semantically very different to regular wait queues
13 * (wait.h). The most important difference is that the simple waitqueue allows
36 * For all the above, note that simple wait queues should _only_ be used under
38 * wait queues in most cases.
92 * returns true if the wait list is not empty
105 * @cond = true; prepare_to_swait_exclusive(&wq_head, &wait, state);
111 * finish_swait(&wq_head, &wait);
115 * observe an empty wait list while the waiter might not observe @cond.
138 * modifications to the wait queue (task_list). in swq_has_sleeper()
[all …]
H A Dumh.h14 #define UMH_NO_WAIT 0x00 /* don't wait at all */
15 #define UMH_WAIT_EXEC 0x01 /* wait for the exec, but not the process */
16 #define UMH_WAIT_PROC 0x02 /* wait for the process to complete */
17 #define UMH_KILLABLE 0x04 /* wait for EXEC/PROC killable */
18 #define UMH_FREEZABLE 0x08 /* wait for EXEC/PROC freezable */
26 int wait; member
34 call_usermodehelper(const char *path, char **argv, char **envp, int wait);
43 call_usermodehelper_exec(struct subprocess_info *info, int wait);
/linux/fs/f2fs/
H A Dgc.h16 #define DEF_GC_THREAD_NOGC_SLEEP_TIME 300000 /* wait 5 min */
64 * will wait on this wait queue.
146 unsigned int *wait) in increase_sleep_time() argument
151 if (*wait == gc_th->no_gc_sleep_time) in increase_sleep_time()
154 if ((long long)*wait + (long long)min_time > (long long)max_time) in increase_sleep_time()
155 *wait = max_time; in increase_sleep_time()
157 *wait += min_time; in increase_sleep_time()
161 unsigned int *wait) in decrease_sleep_time() argument
165 if (*wait == gc_th->no_gc_sleep_time) in decrease_sleep_time()
166 *wait = gc_th->max_sleep_time; in decrease_sleep_time()
[all …]
/linux/net/atm/
H A Dsvc.c13 #include <linux/wait.h>
53 DEFINE_WAIT(wait); in svc_disconnect()
61 prepare_to_wait(sk_sleep(sk), &wait, TASK_UNINTERRUPTIBLE); in svc_disconnect()
66 finish_wait(sk_sleep(sk), &wait); in svc_disconnect()
103 DEFINE_WAIT(wait); in svc_bind()
137 prepare_to_wait(sk_sleep(sk), &wait, TASK_UNINTERRUPTIBLE); in svc_bind()
142 finish_wait(sk_sleep(sk), &wait); in svc_bind()
159 DEFINE_WAIT(wait); in svc_connect()
219 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); in svc_connect()
223 prepare_to_wait(sk_sleep(sk), &wait, in svc_connect()
[all...]
/linux/drivers/scsi/snic/
H A Dvnic_dev.c53 int wait);
246 int wait) in _svnic_dev_cmd2() argument
313 for (delay = 0; delay < wait; delay++) { in _svnic_dev_cmd2()
432 u64 *a0, u64 *a1, int wait) in svnic_dev_cmd() argument
440 err = (*vdev->devcmd_rtn)(vdev, cmd, wait); in svnic_dev_cmd()
452 int wait = VNIC_DVCMD_TMO; in svnic_dev_fw_info() local
465 err = svnic_dev_cmd(vdev, CMD_MCPU_FW_INFO, &a0, &a1, wait); in svnic_dev_fw_info()
477 int wait = VNIC_DVCMD_TMO; in svnic_dev_spec() local
483 err = svnic_dev_cmd(vdev, CMD_DEV_SPEC, &a0, &a1, wait); in svnic_dev_spec()
509 int wait = VNIC_DVCMD_TMO; in svnic_dev_stats_clear() local
[all …]
/linux/Documentation/locking/
H A Dww-mutex-design.rst2 Wound/Wait Deadlock-Proof Mutex Design
5 Please read mutex-design.rst first, as it applies to wait/wound mutexes too.
14 a handful of situations where the driver needs to wait for buffers to
37 and the deadlock handling approach is called Wait-Die. The name is based on
41 and dies. Hence Wait-Die.
42 There is also another algorithm called Wound-Wait:
46 transaction. Hence Wound-Wait.
48 However, the Wound-Wait algorithm is typically stated to generate fewer backoffs
49 compared to Wait-Die, but is, on the other hand, associated with more work than
50 Wait-Die when recovering from a backoff. Wound-Wait is also a preemptive
[all …]
/linux/fs/autofs/
H A Dwaitq.c29 sbi->queues = NULL; /* Erase all wait queues */ in autofs_catatonic_mode()
92 pr_debug("wait id = 0x%08lx, name = %.*s, type=%d\n", in autofs_notify_daemon()
199 * found or NULL to idicate a new wait needs to be created.
202 static int validate_request(struct autofs_wait_queue **wait, in validate_request() argument
214 /* Wait in progress, continue; */ in validate_request()
217 *wait = wq; in validate_request()
221 *wait = NULL; in validate_request()
229 * If we've been asked to wait on an existing expire (NFY_NONE) in validate_request()
230 * but there is no wait in the queue ... in validate_request()
235 * wait or it finished while we waited on the mutex. in validate_request()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_syncobj.c3 * Parts ported from amdgpu (fence wait code).
44 * - Wait for a syncobj's fence to appear and be signaled
51 * - Wait for a given point to appear and/or be signaled
107 * Host-side wait on syncobjs
111 * host-side wait on all of the syncobj fences simultaneously.
112 * If &DRM_SYNCOBJ_WAIT_FLAGS_WAIT_ALL is set, the wait ioctl will wait on
119 * the host-side wait will first wait for the syncobj to receive a non-NULL
120 * fence and then wait on that fence.
124 * to do a host wait in one thread (or process) which waits on GPU work
135 * handles as well as an array of u64 points and does a host-side wait on all
[all …]
/linux/kernel/locking/
H A Dww_mutex.h200 * Wait-Die:
205 * Wound-Wait:
260 * Depending on the algorithm, @a will either need to wait for @b, or die.
304 * Wait-Die; wake a lesser waiter context (when locks held) such that it can
335 * Wound-Wait; wound a lesser @hold_ctx if it holds the lock.
397 * waiting behind us on the wait-list, check if they need to die, or wound us.
402 * This relies on never mixing wait-die/wound-wait on the same wait-list;
405 * The current task must not be on the wait list.
495 * Wound-Wait: If we're wounded, kill ourself.
497 * Wait-Die: If we're trying to acquire a lock already held by an older
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/core/
H A Devent.c98 nvkm_event_ntfy_block_(struct nvkm_event_ntfy *ntfy, bool wait) in nvkm_event_ntfy_block_() argument
102 nvkm_trace(subdev, "event: ntfy block %08x on %d wait:%d\n", ntfy->bits, ntfy->id, wait); in nvkm_event_ntfy_block_()
106 if (wait) in nvkm_event_ntfy_block_()
115 nvkm_event_ntfy_block_(ntfy, ntfy->wait); in nvkm_event_ntfy_block()
125 if (ntfy->wait) in nvkm_event_ntfy_allow()
146 nvkm_event_ntfy_add(struct nvkm_event *event, int id, u32 bits, bool wait, nvkm_event_func func, in nvkm_event_ntfy_add() argument
149 nvkm_trace(event->subdev, "event: ntfy add %08x on %d wait:%d\n", id, bits, wait); in nvkm_event_ntfy_add()
154 ntfy->wait = wait; in nvkm_event_ntfy_add()
159 if (!ntfy->wait) in nvkm_event_ntfy_add()
/linux/fs/squashfs/
H A Ddecompressor_multi.c11 #include <linux/wait.h>
42 wait_queue_head_t wait; member
58 wake_up(&stream->wait); in put_decomp_stream()
75 init_waitqueue_head(&stream->wait); in squashfs_decompressor_create()
145 * let's wait for releasing decomp from other users. in get_decomp_stream()
148 goto wait; in get_decomp_stream()
153 goto wait; in get_decomp_stream()
159 goto wait; in get_decomp_stream()
167 wait: in get_decomp_stream()
174 wait_event(stream->wait, in get_decomp_stream()
/linux/fs/btrfs/
H A Dlocking.h10 #include <linux/wait.h>
94 * Lockdep annotation for wait events.
97 * @lock: The lockdep map corresponding to a wait event
99 * This macro is used to annotate a wait event. In this case a thread acquires
101 * the threads that hold the lock as readers signal the condition for the wait
111 * Protection for the resource/condition of a wait event.
114 * @lock: The lockdep map corresponding to a wait event
116 * Many threads can modify the condition for the wait event at the same time
117 * and signal the threads that block on the wait event. The threads that modify
125 * Used after signaling the condition for a wait event to release the lockdep
[all …]
/linux/include/uapi/drm/
H A Dtegra_drm.h118 * struct drm_tegra_syncpt_wait - parameters for the wait syncpoint IOCTL
124 * ID of the syncpoint to wait on.
131 * Threshold value for which to wait.
138 * Timeout, in milliseconds, to wait.
145 * The new syncpoint value after the wait. Set by the kernel upon
223 * struct drm_tegra_get_syncpt_base - parameters for the get wait base IOCTL
230 * wait base.
237 * ID of the syncpoint for which to obtain the wait base.
244 * The ID of the wait base corresponding to the client syncpoint. Set
357 * struct drm_tegra_waitchk - wait check structure
[all …]

12345678910>>...155