/linux/drivers/infiniband/hw/hfi1/ |
H A D | iowait.h | 102 struct iowait_work *wait, 107 void (*wakeup)(struct iowait *wait, int reason); 108 void (*sdma_drained)(struct iowait *wait); 109 void (*init_priority)(struct iowait *wait); 121 struct iowait_work wait[IOWAIT_SES]; member 126 void iowait_set_flag(struct iowait *wait, u32 flag); 127 bool iowait_flag_set(struct iowait *wait, u32 flag); 128 void iowait_clear_flag(struct iowait *wait, u32 flag); 130 void iowait_init(struct iowait *wait, u32 tx_limit, 134 struct iowait_work *wait, [all …]
|
H A D | iowait.c | 12 void iowait_set_flag(struct iowait *wait, u32 flag) in iowait_set_flag() argument 14 trace_hfi1_iowait_set(wait, flag); in iowait_set_flag() 15 set_bit(flag, &wait->flags); in iowait_set_flag() 18 bool iowait_flag_set(struct iowait *wait, u32 flag) in iowait_flag_set() argument 20 return test_bit(flag, &wait->flags); in iowait_flag_set() 23 inline void iowait_clear_flag(struct iowait *wait, u32 flag) in iowait_clear_flag() argument 25 trace_hfi1_iowait_clear(wait, flag); in iowait_clear_flag() 26 clear_bit(flag, &wait->flags); in iowait_clear_flag() 41 void iowait_init(struct iowait *wait, u32 tx_limit, in iowait_init() argument 45 struct iowait_work *wait, in iowait_init() argument [all …]
|
H A D | trace_iowait.h | 17 TP_PROTO(struct iowait *wait, u32 flag), 18 TP_ARGS(wait, flag), 26 __entry->addr = (unsigned long)wait; 27 __entry->flags = wait->flags; 29 __entry->qpn = iowait_to_qp(wait)->ibqp.qp_num; 41 TP_PROTO(struct iowait *wait, u32 flag), 42 TP_ARGS(wait, flag)); 45 TP_PROTO(struct iowait *wait, u32 flag), 46 TP_ARGS(wait, flag));
|
H A D | vnic_sdma.c | 155 ret = sdma_send_txreq(sde, iowait_get_ib_work(&vnic_sdma->wait), in hfi1_vnic_send_dma() 163 iowait_starve_clear(vnic_sdma->pkts_sent, &vnic_sdma->wait); in hfi1_vnic_send_dma() 187 struct iowait_work *wait, in hfi1_vnic_sdma_sleep() argument 193 container_of(wait->iow, struct hfi1_vnic_sdma, wait); in hfi1_vnic_sdma_sleep() 202 if (list_empty(&vnic_sdma->wait.list)) { in hfi1_vnic_sdma_sleep() 203 iowait_get_priority(wait->iow); in hfi1_vnic_sdma_sleep() 204 iowait_queue(pkts_sent, wait->iow, &sde->dmawait); in hfi1_vnic_sdma_sleep() 217 static void hfi1_vnic_sdma_wakeup(struct iowait *wait, int reason) in hfi1_vnic_sdma_wakeup() argument 220 container_of(wait, struct hfi1_vnic_sdma, wait); in hfi1_vnic_sdma_wakeup() 243 iowait_init(&vnic_sdma->wait, 0, NULL, NULL, in hfi1_vnic_sdma_init() [all …]
|
/linux/drivers/gpu/drm/i915/display/ |
H A D | intel_display_rps.c | 15 struct wait_queue_entry wait; member 24 struct wait_rps_boost *wait = container_of(_wait, typeof(*wait), wait); in do_rps_boost() local 25 struct i915_request *rq = wait->request; in do_rps_boost() 36 drm_crtc_vblank_put(wait->crtc); in do_rps_boost() 38 list_del(&wait->wait.entry); in do_rps_boost() 39 kfree(wait); in do_rps_boost() 46 struct wait_rps_boost *wait; in intel_display_rps_boost_after_vblank() local 57 wait = kmalloc(sizeof(*wait), GFP_KERNEL); in intel_display_rps_boost_after_vblank() 58 if (!wait) { in intel_display_rps_boost_after_vblank() 63 wait->request = to_request(dma_fence_get(fence)); in intel_display_rps_boost_after_vblank() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvif/ |
H A D | timer.c | 26 nvif_timer_wait_test(struct nvif_timer_wait *wait) in nvif_timer_wait_test() argument 28 u64 time = nvif_device_time(wait->device); in nvif_timer_wait_test() 30 if (wait->reads == 0) { in nvif_timer_wait_test() 31 wait->time0 = time; in nvif_timer_wait_test() 32 wait->time1 = time; in nvif_timer_wait_test() 35 if (wait->time1 == time) { in nvif_timer_wait_test() 36 if (WARN_ON(wait->reads++ == 16)) in nvif_timer_wait_test() 39 wait->time1 = time; in nvif_timer_wait_test() 40 wait->reads = 1; in nvif_timer_wait_test() 43 if (wait->time1 - wait->time0 > wait->limit) in nvif_timer_wait_test() [all …]
|
/linux/fs/bcachefs/ |
H A D | clock.c | 62 struct io_clock_wait *wait = container_of(timer, in io_clock_wait_fn() local 65 wait->expired = 1; in io_clock_wait_fn() 66 wake_up_process(wait->task); in io_clock_wait_fn() 71 struct io_clock_wait *wait = container_of(timer, in io_clock_cpu_timeout() local 74 wait->expired = 1; in io_clock_cpu_timeout() 75 wake_up_process(wait->task); in io_clock_cpu_timeout() 80 struct io_clock_wait wait; in bch2_io_clock_schedule_timeout() local 83 wait.io_timer.expire = until; in bch2_io_clock_schedule_timeout() 84 wait.io_timer.fn = io_clock_wait_fn; in bch2_io_clock_schedule_timeout() 85 wait.task = current; in bch2_io_clock_schedule_timeout() [all …]
|
H A D | thread_with_file.c | 95 init_waitqueue_head(&buf->wait); in stdio_buf_init() 105 wake_up(&thr->stdio.input.wait); in thread_with_stdio_done() 106 wake_up(&thr->stdio.output.wait); in thread_with_stdio_done() 119 ret = wait_event_interruptible(buf->wait, stdio_redirect_has_output(&thr->stdio)); in thread_with_stdio_read() 198 wake_up(&buf->wait); in thread_with_stdio_write() 205 ret = wait_event_interruptible(buf->wait, in thread_with_stdio_write() 215 static __poll_t thread_with_stdio_poll(struct file *file, struct poll_table_struct *wait) in thread_with_stdio_poll() argument 220 poll_wait(file, &thr->stdio.output.wait, wait); in thread_with_stdio_poll() 221 poll_wait(file, &thr->stdio.input.wait, wait); in thread_with_stdio_poll() 234 static __poll_t thread_with_stdout_poll(struct file *file, struct poll_table_struct *wait) in thread_with_stdout_poll() argument [all …]
|
H A D | six.h | 211 struct six_lock_waiter *wait, 230 struct six_lock_waiter *wait, in six_lock_waiter() argument 233 return six_lock_ip_waiter(lock, type, wait, should_sleep_fn, p, _THIS_IP_); in six_lock_waiter() 251 struct six_lock_waiter wait; in six_lock_ip() local 253 return six_lock_ip_waiter(lock, type, &wait, should_sleep_fn, p, ip); in six_lock_ip() 269 struct six_lock_waiter wait; in six_lock_type() local 271 return six_lock_ip_waiter(lock, type, &wait, should_sleep_fn, p, _THIS_IP_); in six_lock_type() 325 struct six_lock_waiter *wait, \ 329 return six_lock_ip_waiter(lock, SIX_LOCK_##type, wait, should_sleep_fn, p, ip);\
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/timer/ |
H A D | base.c | 27 nvkm_timer_wait_test(struct nvkm_timer_wait *wait) in nvkm_timer_wait_test() argument 29 struct nvkm_subdev *subdev = &wait->tmr->subdev; in nvkm_timer_wait_test() 30 u64 time = nvkm_timer_read(wait->tmr); in nvkm_timer_wait_test() 32 if (wait->reads == 0) { in nvkm_timer_wait_test() 33 wait->time0 = time; in nvkm_timer_wait_test() 34 wait->time1 = time; in nvkm_timer_wait_test() 37 if (wait->time1 == time) { in nvkm_timer_wait_test() 38 if (wait->reads++ == 16) { in nvkm_timer_wait_test() 43 wait->time1 = time; in nvkm_timer_wait_test() 44 wait->reads = 1; in nvkm_timer_wait_test() [all …]
|
/linux/fs/ |
H A D | fs_pin.c | 16 spin_lock_irq(&pin->wait.lock); in pin_remove() 18 wake_up_locked(&pin->wait); in pin_remove() 19 spin_unlock_irq(&pin->wait.lock); in pin_remove() 32 wait_queue_entry_t wait; in pin_kill() local 38 init_wait(&wait); in pin_kill() 39 spin_lock_irq(&p->wait.lock); in pin_kill() 42 spin_unlock_irq(&p->wait.lock); in pin_kill() 48 spin_unlock_irq(&p->wait.lock); in pin_kill() 52 __add_wait_queue(&p->wait, &wait); in pin_kill() 55 spin_unlock_irq(&p->wait.lock); in pin_kill() [all …]
|
/linux/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_dev.c | 213 int wait) in _vnic_dev_cmd() argument 242 for (delay = 0; delay < wait; delay++) { in _vnic_dev_cmd() 281 int wait) in _vnic_dev_cmd2() argument 332 for (delay = 0; delay < wait; delay++) { in _vnic_dev_cmd2() 440 u64 *a0, u64 *a1, int wait) in vnic_dev_cmd_proxy() argument 452 err = vdev->devcmd_rtn(vdev, proxy_cmd, wait); in vnic_dev_cmd_proxy() 473 enum vnic_devcmd_cmd cmd, u64 *a0, u64 *a1, int wait) in vnic_dev_cmd_no_proxy() argument 480 err = vdev->devcmd_rtn(vdev, cmd, wait); in vnic_dev_cmd_no_proxy() 501 u64 *a0, u64 *a1, int wait) in vnic_dev_cmd() argument 508 a0, a1, wait); in vnic_dev_cmd() [all …]
|
/linux/kernel/sched/ |
H A D | completion.c | 20 raw_spin_lock_irqsave(&x->wait.lock, flags); in complete_with_flags() 24 swake_up_locked(&x->wait, wake_flags); in complete_with_flags() 25 raw_spin_unlock_irqrestore(&x->wait.lock, flags); in complete_with_flags() 73 raw_spin_lock_irqsave(&x->wait.lock, flags); in complete_all() 75 swake_up_all_locked(&x->wait); in complete_all() 76 raw_spin_unlock_irqrestore(&x->wait.lock, flags); in complete_all() 85 DECLARE_SWAITQUEUE(wait); in do_wait_for_common() 92 __prepare_to_swait(&x->wait, &wait); in do_wait_for_common() 94 raw_spin_unlock_irq(&x->wait.lock); in do_wait_for_common() 96 raw_spin_lock_irq(&x->wait.lock); in do_wait_for_common() [all …]
|
H A D | swait.c | 84 void __prepare_to_swait(struct swait_queue_head *q, struct swait_queue *wait) in __prepare_to_swait() argument 86 wait->task = current; in __prepare_to_swait() 87 if (list_empty(&wait->task_list)) in __prepare_to_swait() 88 list_add_tail(&wait->task_list, &q->task_list); in __prepare_to_swait() 91 void prepare_to_swait_exclusive(struct swait_queue_head *q, struct swait_queue *wait, int state) in prepare_to_swait_exclusive() argument 96 __prepare_to_swait(q, wait); in prepare_to_swait_exclusive() 102 long prepare_to_swait_event(struct swait_queue_head *q, struct swait_queue *wait, int state) in prepare_to_swait_event() argument 113 list_del_init(&wait->task_list); in prepare_to_swait_event() 116 __prepare_to_swait(q, wait); in prepare_to_swait_event() 125 void __finish_swait(struct swait_queue_head *q, struct swait_queue *wait) in __finish_swait() argument [all …]
|
/linux/include/linux/ |
H A D | smp.h | 51 int wait); 54 void *info, bool wait, const struct cpumask *mask); 69 static inline void on_each_cpu(smp_call_func_t func, void *info, int wait) in on_each_cpu() argument 71 on_each_cpu_cond_mask(NULL, func, info, wait, cpu_online_mask); in on_each_cpu() 91 smp_call_func_t func, void *info, bool wait) in on_each_cpu_mask() argument 93 on_each_cpu_cond_mask(NULL, func, info, wait, mask); in on_each_cpu_mask() 103 smp_call_func_t func, void *info, bool wait) in on_each_cpu_cond() argument 105 on_each_cpu_cond_mask(cond_func, func, info, wait, cpu_online_mask); in on_each_cpu_cond() 162 void smp_call_function(smp_call_func_t func, void *info, int wait); 164 smp_call_func_t func, void *info, bool wait); [all …]
|
/linux/drivers/gpu/drm/omapdrm/ |
H A D | omap_irq.c | 22 struct omap_irq_wait *wait; in omap_irq_update() local 27 list_for_each_entry(wait, &priv->wait_list, node) in omap_irq_update() 28 irqmask |= wait->irqmask; in omap_irq_update() 35 static void omap_irq_wait_handler(struct omap_irq_wait *wait) in omap_irq_wait_handler() argument 37 wait->count--; in omap_irq_wait_handler() 38 wake_up(&wait->wq); in omap_irq_wait_handler() 45 struct omap_irq_wait *wait = kzalloc(sizeof(*wait), GFP_KERNEL); in omap_irq_wait_init() local 48 init_waitqueue_head(&wait->wq); in omap_irq_wait_init() 49 wait->irqmask = irqmask; in omap_irq_wait_init() 50 wait->count = count; in omap_irq_wait_init() [all …]
|
/linux/drivers/scsi/fnic/ |
H A D | vnic_dev.c | 62 int wait); 249 static int vnic_dev_cmd1(struct vnic_dev *vdev, enum vnic_devcmd_cmd cmd, int wait) in vnic_dev_cmd1() argument 283 for (delay = 0; delay < wait; delay++) { in vnic_dev_cmd1() 312 int wait) in vnic_dev_cmd2() argument 379 for (delay = 0; delay < wait; delay++) { in vnic_dev_cmd2() 501 enum vnic_devcmd_cmd cmd, u64 *a0, u64 *a1, int wait) in vnic_dev_cmd_no_proxy() argument 508 err = (*vdev->devcmd_rtn)(vdev, cmd, wait); in vnic_dev_cmd_no_proxy() 518 u64 *a0, u64 *a1, int wait) in vnic_dev_cmd() argument 525 return vnic_dev_cmd_no_proxy(vdev, cmd, a0, a1, wait); in vnic_dev_cmd() 534 int wait = 1000; in vnic_dev_fw_info() local [all …]
|
/linux/fs/f2fs/ |
H A D | gc.h | 16 #define DEF_GC_THREAD_NOGC_SLEEP_TIME 300000 /* wait 5 min */ 52 * will wait on this wait queue. 127 unsigned int *wait) in increase_sleep_time() 132 if (*wait == gc_th->no_gc_sleep_time) in increase_sleep_time() 135 if ((long long)*wait + (long long)min_time > (long long)max_time) in increase_sleep_time() 136 *wait = max_time; in increase_sleep_time() 138 *wait += min_time; in increase_sleep_time() 142 unsigned int *wait) in decrease_sleep_time() 146 if (*wait in decrease_sleep_time() 126 increase_sleep_time(struct f2fs_gc_kthread * gc_th,unsigned int * wait) increase_sleep_time() argument 141 decrease_sleep_time(struct f2fs_gc_kthread * gc_th,unsigned int * wait) decrease_sleep_time() argument [all...] |
/linux/net/core/ |
H A D | stream.c | 41 wake_up_interruptible_poll(&wq->wait, EPOLLOUT | in sk_stream_write_space() 58 DEFINE_WAIT_FUNC(wait, woken_wake_function); in sk_stream_wait_connect() 73 add_wait_queue(sk_sleep(sk), &wait); in sk_stream_wait_connect() 78 ~(TCPF_ESTABLISHED | TCPF_CLOSE_WAIT)), &wait); in sk_stream_wait_connect() 79 remove_wait_queue(sk_sleep(sk), &wait); in sk_stream_wait_connect() 99 DEFINE_WAIT_FUNC(wait, woken_wake_function); in sk_stream_wait_close() 101 add_wait_queue(sk_sleep(sk), &wait); in sk_stream_wait_close() 104 if (sk_wait_event(sk, &timeout, !sk_stream_closing(sk), &wait)) in sk_stream_wait_close() 108 remove_wait_queue(sk_sleep(sk), &wait); in sk_stream_wait_close() 123 DEFINE_WAIT_FUNC(wait, woken_wake_function); in sk_stream_wait_memory() [all …]
|
/linux/kernel/ |
H A D | umh.c | 119 if (!(sub_info->wait & UMH_WAIT_PROC)) in call_usermodehelper_exec_async() 163 if (sub_info->wait & UMH_WAIT_PROC) { in call_usermodehelper_exec_work() 212 DEFINE_WAIT(wait); in usermodehelper_read_trylock() 217 prepare_to_wait(&usermodehelper_disabled_waitq, &wait, in usermodehelper_read_trylock() 235 finish_wait(&usermodehelper_disabled_waitq, &wait); in usermodehelper_read_trylock() 242 DEFINE_WAIT(wait); in usermodehelper_read_lock_wait() 249 prepare_to_wait(&usermodehelper_disabled_waitq, &wait, in usermodehelper_read_lock_wait() 262 finish_wait(&usermodehelper_disabled_waitq, &wait); in usermodehelper_read_lock_wait() 402 int call_usermodehelper_exec(struct subprocess_info *sub_info, int wait) in call_usermodehelper_exec() argument 431 sub_info->complete = (wait == UMH_NO_WAIT) ? NULL : &done; in call_usermodehelper_exec() [all …]
|
/linux/drivers/scsi/snic/ |
H A D | vnic_dev.c | 55 int wait); 248 int wait) in _svnic_dev_cmd2() argument 315 for (delay = 0; delay < wait; delay++) { in _svnic_dev_cmd2() 434 u64 *a0, u64 *a1, int wait) in svnic_dev_cmd() argument 442 err = (*vdev->devcmd_rtn)(vdev, cmd, wait); in svnic_dev_cmd() 454 int wait = VNIC_DVCMD_TMO; in svnic_dev_fw_info() local 467 err = svnic_dev_cmd(vdev, CMD_MCPU_FW_INFO, &a0, &a1, wait); in svnic_dev_fw_info() 479 int wait = VNIC_DVCMD_TMO; in svnic_dev_spec() local 485 err = svnic_dev_cmd(vdev, CMD_DEV_SPEC, &a0, &a1, wait); in svnic_dev_spec() 511 int wait = VNIC_DVCMD_TMO; in svnic_dev_stats_clear() local [all …]
|
/linux/net/atm/ |
H A D | svc.c | 53 DEFINE_WAIT(wait); in svc_disconnect() 61 prepare_to_wait(sk_sleep(sk), &wait, TASK_UNINTERRUPTIBLE); in svc_disconnect() 66 finish_wait(sk_sleep(sk), &wait); in svc_disconnect() 103 DEFINE_WAIT(wait); in svc_bind() 137 prepare_to_wait(sk_sleep(sk), &wait, TASK_UNINTERRUPTIBLE); in svc_bind() 142 finish_wait(sk_sleep(sk), &wait); in svc_bind() 159 DEFINE_WAIT(wait); in svc_connect() 219 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); in svc_connect() 223 prepare_to_wait(sk_sleep(sk), &wait, in svc_connect() 242 prepare_to_wait(sk_sleep(sk), &wait, in svc_connect() [all …]
|
/linux/kernel/module/ |
H A D | kmod.c | 72 static int call_modprobe(char *orig_module_name, int wait) in call_modprobe() argument 103 ret = call_usermodehelper_exec(info, wait | UMH_KILLABLE); in call_modprobe() 132 int __request_module(bool wait, const char *fmt, ...) in __request_module() argument 144 WARN_ON_ONCE(wait && current_is_async()); in __request_module() 166 trace_module_request(module_name, wait, _RET_IP_); in __request_module() 168 if (kmod_dup_request_exists_wait(module_name, wait, &dup_ret)) { in __request_module() 173 ret = call_modprobe(module_name, wait ? UMH_WAIT_PROC : UMH_WAIT_EXEC); in __request_module()
|
/linux/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | event.c | 98 nvkm_event_ntfy_block_(struct nvkm_event_ntfy *ntfy, bool wait) in nvkm_event_ntfy_block_() argument 102 nvkm_trace(subdev, "event: ntfy block %08x on %d wait:%d\n", ntfy->bits, ntfy->id, wait); in nvkm_event_ntfy_block_() 106 if (wait) in nvkm_event_ntfy_block_() 115 nvkm_event_ntfy_block_(ntfy, ntfy->wait); in nvkm_event_ntfy_block() 125 if (ntfy->wait) in nvkm_event_ntfy_allow() 146 nvkm_event_ntfy_add(struct nvkm_event *event, int id, u32 bits, bool wait, nvkm_event_func func, in nvkm_event_ntfy_add() argument 149 nvkm_trace(event->subdev, "event: ntfy add %08x on %d wait:%d\n", id, bits, wait); in nvkm_event_ntfy_add() 154 ntfy->wait = wait; in nvkm_event_ntfy_add() 159 if (!ntfy->wait) in nvkm_event_ntfy_add()
|
/linux/fs/squashfs/ |
H A D | decompressor_multi.c | 42 wait_queue_head_t wait; member 58 wake_up(&stream->wait); in put_decomp_stream() 75 init_waitqueue_head(&stream->wait); in squashfs_decompressor_create() 148 goto wait; in get_decomp_stream() 153 goto wait; in get_decomp_stream() 159 goto wait; in get_decomp_stream() 167 wait: in get_decomp_stream() 174 wait_event(stream->wait, in get_decomp_stream()
|