| /linux/net/x25/ |
| H A D | x25_in.c | 213 int queued = 0; in x25_state3_machine() 280 queued = 1; in x25_state3_machine() 318 queued = !sock_queue_rcv_skb(sk, skb); in x25_state3_machine() 322 queued = 1; in x25_state3_machine() 333 return queued; in x25_state3_machine() 421 int queued = 0, frametype, ns, nr, q, d, m; in x25_process_rx_frame() 430 queued = x25_state1_machine(sk, skb, frametype); in x25_process_rx_frame() 433 queued = x25_state2_machine(sk, skb, frametype); in x25_process_rx_frame() 436 queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in x25_process_rx_frame() 439 queued in x25_process_rx_frame() 210 int queued = 0; x25_state3_machine() local 418 int queued = 0, frametype, ns, nr, q, d, m; x25_process_rx_frame() local 450 int queued = x25_process_rx_frame(sk, skb); x25_backlog_rcv() local [all...] |
| H A D | x25_dev.c | 51 int queued = 1; in x25_receive_data() local 56 queued = x25_process_rx_frame(sk, skb); in x25_receive_data() 58 queued = !sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)); in x25_receive_data() 62 return queued; in x25_receive_data()
|
| /linux/net/rose/ |
| H A D | rose_in.c | 106 int queued = 0; in rose_state3_machine() local 169 queued = 1; in rose_state3_machine() 206 return queued; in rose_state3_machine() 267 int queued = 0, frametype, ns, nr, q, d, m; in rose_process_rx_frame() local 276 queued = rose_state1_machine(sk, skb, frametype); in rose_process_rx_frame() 279 queued = rose_state2_machine(sk, skb, frametype); in rose_process_rx_frame() 282 queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in rose_process_rx_frame() 285 queued = rose_state4_machine(sk, skb, frametype); in rose_process_rx_frame() 288 queued = rose_state5_machine(sk, skb, frametype); in rose_process_rx_frame() 294 return queued; in rose_process_rx_frame()
|
| /linux/net/dccp/ |
| H A D | input.c | |
| /linux/net/ax25/ |
| H A D | ax25_std_in.c | 143 int queued = 0; in ax25_std_state3_machine() local 225 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state3_machine() 258 return queued; in ax25_std_state3_machine() 268 int queued = 0; in ax25_std_state4_machine() local 380 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state4_machine() 413 return queued; in ax25_std_state4_machine() 421 int queued = 0, frametype, ns, nr, pf; in ax25_std_frame_in() local 427 queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 430 queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 433 queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_std_frame_in() [all …]
|
| H A D | ax25_ds_in.c | 147 int queued = 0; in ax25_ds_state3_machine() local 240 queued = ax25_rx_iframe(ax25, skb); in ax25_ds_state3_machine() 273 return queued; in ax25_ds_state3_machine() 281 int queued = 0, frametype, ns, nr, pf; in ax25_ds_frame_in() local 287 queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 290 queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 293 queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_ds_frame_in() 297 return queued; in ax25_ds_frame_in()
|
| H A D | ax25_in.c | 103 int queued = 0; in ax25_rx_iframe() local 145 queued = 1; in ax25_rx_iframe() 151 return queued; in ax25_rx_iframe() 159 int queued = 0; in ax25_process_rx_frame() local 167 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame() 173 queued = ax25_ds_frame_in(ax25, skb, type); in ax25_process_rx_frame() 175 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame() 180 return queued; in ax25_process_rx_frame()
|
| /linux/net/netrom/ |
| H A D | nr_in.c | 153 int queued = 0; in nr_state3_machine() local 225 queued = 1; in nr_state3_machine() 272 return queued; in nr_state3_machine() 279 int queued = 0, frametype; in nr_process_rx_frame() local 288 queued = nr_state1_machine(sk, skb, frametype); in nr_process_rx_frame() 291 queued = nr_state2_machine(sk, skb, frametype); in nr_process_rx_frame() 294 queued = nr_state3_machine(sk, skb, frametype); in nr_process_rx_frame() 300 return queued; in nr_process_rx_frame()
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_flip_work.c | 52 list_add_tail(&task->node, &work->queued); in drm_flip_work_queue_task() 95 list_splice_tail(&work->queued, &work->commited); in drm_flip_work_commit() 96 INIT_LIST_HEAD(&work->queued); in drm_flip_work_commit() 139 INIT_LIST_HEAD(&work->queued); in drm_flip_work_init() 156 WARN_ON(!list_empty(&work->queued) || !list_empty(&work->commited)); in drm_flip_work_cleanup()
|
| H A D | drm_pagemap_util.c | 39 struct completion queued; member 102 init_completion(&cache->queued); in drm_pagemap_cache_create_devm() 194 if (!try_wait_for_completion(&cache->queued)) { in drm_pagemap_get_from_cache() 196 err = wait_for_completion_interruptible(&cache->queued); in drm_pagemap_get_from_cache() 237 reinit_completion(&cache->queued); in drm_pagemap_cache_set_pagemap() 342 complete_all(&cache->queued); in drm_pagemap_shrinker_add()
|
| /linux/tools/testing/selftests/net/mptcp/ |
| H A D | mptcp_inq.c | 218 int nsd, ret, queued = -1; in wait_for_ack() local 221 ret = ioctl(fd, TIOCOUTQ, &queued); in wait_for_ack() 229 if ((size_t)queued > total) in wait_for_ack() 230 xerror("TIOCOUTQ %u, but only %zu expected\n", queued, total); in wait_for_ack() 231 assert(nsd <= queued); in wait_for_ack() 233 if (queued == 0) in wait_for_ack() 366 unsigned int queued; in process_one_client() local 368 ret = ioctl(fd, FIONREAD, &queued); in process_one_client() 371 if (queued > expect_len) in process_one_client() 373 queued, expect_len); in process_one_client() [all …]
|
| /linux/Documentation/userspace-api/media/mediactl/ |
| H A D | media-request-ioc-queue.rst | 34 If the request was successfully queued, then the file descriptor can be 37 If the request was already queued before, then ``EBUSY`` is returned. 42 Once a request is queued, then the driver is required to gracefully handle 49 queued directly and you next try to queue a request, or vice versa. 62 The request was already queued or the application queued the first
|
| H A D | media-request-ioc-reinit.rst | 40 A request can only be re-initialized if it either has not been queued 41 yet, or if it was queued and completed. Otherwise it will set ``errno`` 51 The request is queued but not yet completed.
|
| H A D | request-api.rst | 24 buffer queues since in practice only one buffer would be queued at a time. 59 instead of being immediately applied, and buffers queued to a request do not 60 enter the regular buffer queue until the request itself is queued. 66 queued by calling :ref:`MEDIA_REQUEST_IOC_QUEUE` on the request file descriptor. 68 A queued request cannot be modified anymore. 86 a buffer was queued via a request or vice versa will result in an ``EBUSY`` 109 request that has been queued but not yet completed will return ``EBUSY`` 121 longer in use by the kernel. That is, if the request is queued and then the 165 Once the request is fully prepared, it can be queued to the driver: 245 Once the request is fully prepared, it can be queued to the driver:
|
| /linux/drivers/md/ |
| H A D | dm-cache-background-tracker.c | 21 struct list_head queued; member 42 INIT_LIST_HEAD(&b->queued); in btracker_create() 55 list_for_each_entry_safe (w, tmp, &b->queued, list) { in btracker_destroy() 195 list_add(&w->list, &b->queued); in btracker_queue() 209 if (list_empty(&b->queued)) in btracker_issue() 212 w = list_first_entry(&b->queued, struct bt_work, list); in btracker_issue()
|
| /linux/security/integrity/ima/ |
| H A D | ima_asymmetric_keys.c | 33 bool queued = false; in ima_post_key_create_or_update() local 43 queued = ima_queue_key(keyring, payload, payload_len); in ima_post_key_create_or_update() 45 if (queued) in ima_post_key_create_or_update()
|
| H A D | ima_queue_keys.c | 107 bool queued = false; in ima_queue_key() local 117 queued = true; in ima_queue_key() 121 if (!queued) in ima_queue_key() 124 return queued; in ima_queue_key()
|
| /linux/virt/kvm/ |
| H A D | async_pf.c | 111 * need to be flushed (but sanity check that the work wasn't queued). in kvm_flush_and_free_async_pf_work() 150 vcpu->async_pf.queued = 0; in kvm_clear_async_pf_completion_queue() 170 vcpu->async_pf.queued--; in kvm_check_async_pf_completion() 184 if (vcpu->async_pf.queued >= ASYNC_PF_PER_VCPU) in kvm_setup_async_pf() 208 vcpu->async_pf.queued++; in kvm_setup_async_pf() 239 vcpu->async_pf.queued++; in kvm_async_pf_wakeup_all()
|
| /linux/fs/xfs/ |
| H A D | xfs_mru_cache.c | 102 unsigned int queued; /* work has been queued */ member 204 if (!mru->queued) { in _xfs_mru_cache_list_insert() 205 mru->queued = 1; in _xfs_mru_cache_list_insert() 279 mru->queued = next; in _xfs_mru_cache_reap() 280 if ((mru->queued > 0)) { in _xfs_mru_cache_reap() 382 if (mru->queued) { in xfs_mru_cache_flush()
|
| /linux/drivers/media/platform/renesas/vsp1/ |
| H A D | vsp1_dl.c | 229 struct vsp1_dl_list *queued; member 860 if (!dlm->queued) in vsp1_dl_list_hw_update_pending() 918 __vsp1_dl_list_put(dlm->queued); in vsp1_dl_list_commit_continuous() 919 dlm->queued = dl; in vsp1_dl_list_commit_continuous() 1039 if (dlm->queued) { in vsp1_dlm_irq_frame_end() 1040 if (dlm->queued->flags & VSP1_DL_FRAME_END_INTERNAL) in vsp1_dlm_irq_frame_end() 1042 dlm->queued->flags &= ~VSP1_DL_FRAME_END_INTERNAL; in vsp1_dlm_irq_frame_end() 1045 dlm->active = dlm->queued; in vsp1_dlm_irq_frame_end() 1046 dlm->queued = NULL; in vsp1_dlm_irq_frame_end() 1057 dlm->queued = dlm->pending; in vsp1_dlm_irq_frame_end() [all …]
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | null.c | 123 int queued; in ublk_null_queue_io() 126 queued = null_queue_auto_zc_io(t, q, tag); in ublk_null_queue_io() 128 queued = null_queue_zc_io(t, q, tag); in ublk_null_queue_io() 133 ublk_queued_tgt_io(t, q, tag, queued); in ublk_null_queue_io() 121 int queued; ublk_null_queue_io() local
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | vidioc-streamon.rst | 51 If ``VIDIOC_STREAMON`` fails then any already queued buffers will remain 52 queued. 63 If buffers have been queued with :ref:`VIDIOC_QBUF` and 65 ``VIDIOC_STREAMON``, then those queued buffers will also be removed from 77 but ``VIDIOC_STREAMOFF`` will return queued buffers to their starting
|
| /linux/Documentation/usb/ |
| H A D | ohci.rst | 22 - interrupt transfers can be larger, and can be queued 28 types can be queued. That was also true in "usb-ohci", except for interrupt 30 to overhead in IRQ processing. When interrupt transfers are queued, those
|
| /linux/Documentation/features/locking/queued-rwlocks/ |
| H A D | arch-support.txt | 2 # Feature name: queued-rwlocks 4 # description: arch supports queued rwlocks
|
| /linux/Documentation/features/locking/queued-spinlocks/ |
| H A D | arch-support.txt | 2 # Feature name: queued-spinlocks 4 # description: arch supports queued spinlocks
|