| /linux/net/x25/ |
| H A D | x25_in.c | 210 int queued = 0; in x25_state3_machine() local 277 queued = 1; in x25_state3_machine() 315 queued = !sock_queue_rcv_skb(sk, skb); in x25_state3_machine() 319 queued = 1; in x25_state3_machine() 330 return queued; in x25_state3_machine() 418 int queued = 0, frametype, ns, nr, q, d, m; in x25_process_rx_frame() local 427 queued = x25_state1_machine(sk, skb, frametype); in x25_process_rx_frame() 430 queued = x25_state2_machine(sk, skb, frametype); in x25_process_rx_frame() 433 queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in x25_process_rx_frame() 436 queued = x25_state4_machine(sk, skb, frametype); in x25_process_rx_frame() [all …]
|
| H A D | x25_dev.c | 51 int queued = 1; in x25_receive_data() local 56 queued = x25_process_rx_frame(sk, skb); in x25_receive_data() 58 queued = !sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)); in x25_receive_data() 62 return queued; in x25_receive_data()
|
| /linux/net/rose/ |
| H A D | rose_in.c | 106 int queued = 0; in rose_state3_machine() local 169 queued = 1; in rose_state3_machine() 206 return queued; in rose_state3_machine() 267 int queued = 0, frametype, ns, nr, q, d, m; in rose_process_rx_frame() local 276 queued = rose_state1_machine(sk, skb, frametype); in rose_process_rx_frame() 279 queued = rose_state2_machine(sk, skb, frametype); in rose_process_rx_frame() 282 queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in rose_process_rx_frame() 285 queued = rose_state4_machine(sk, skb, frametype); in rose_process_rx_frame() 288 queued = rose_state5_machine(sk, skb, frametype); in rose_process_rx_frame() 294 return queued; in rose_process_rx_frame()
|
| /linux/net/dccp/ |
| H A D | input.c | |
| /linux/net/ax25/ |
| H A D | ax25_std_in.c | 143 int queued = 0; in ax25_std_state3_machine() local 225 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state3_machine() 258 return queued; in ax25_std_state3_machine() 268 int queued = 0; in ax25_std_state4_machine() local 380 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state4_machine() 413 return queued; in ax25_std_state4_machine() 421 int queued = 0, frametype, ns, nr, pf; in ax25_std_frame_in() local 427 queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 430 queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 433 queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_std_frame_in() [all …]
|
| H A D | ax25_ds_in.c | 147 int queued = 0; in ax25_ds_state3_machine() local 240 queued = ax25_rx_iframe(ax25, skb); in ax25_ds_state3_machine() 273 return queued; in ax25_ds_state3_machine() 281 int queued = 0, frametype, ns, nr, pf; in ax25_ds_frame_in() local 287 queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 290 queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 293 queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_ds_frame_in() 297 return queued; in ax25_ds_frame_in()
|
| H A D | ax25_in.c | 103 int queued = 0; in ax25_rx_iframe() local 145 queued = 1; in ax25_rx_iframe() 151 return queued; in ax25_rx_iframe() 159 int queued = 0; in ax25_process_rx_frame() local 167 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame() 173 queued = ax25_ds_frame_in(ax25, skb, type); in ax25_process_rx_frame() 175 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame() 180 return queued; in ax25_process_rx_frame()
|
| /linux/net/netrom/ |
| H A D | nr_in.c | 153 int queued = 0; in nr_state3_machine() local 225 queued = 1; in nr_state3_machine() 272 return queued; in nr_state3_machine() 279 int queued = 0, frametype; in nr_process_rx_frame() local 288 queued = nr_state1_machine(sk, skb, frametype); in nr_process_rx_frame() 291 queued = nr_state2_machine(sk, skb, frametype); in nr_process_rx_frame() 294 queued = nr_state3_machine(sk, skb, frametype); in nr_process_rx_frame() 300 return queued; in nr_process_rx_frame()
|
| /linux/tools/testing/selftests/net/mptcp/ |
| H A D | mptcp_inq.c | 217 int nsd, ret, queued = -1; in wait_for_ack() local 220 ret = ioctl(fd, TIOCOUTQ, &queued); in wait_for_ack() 228 if ((size_t)queued > total) in wait_for_ack() 229 xerror("TIOCOUTQ %u, but only %zu expected\n", queued, total); in wait_for_ack() 230 assert(nsd <= queued); in wait_for_ack() 232 if (queued == 0) in wait_for_ack() 241 xerror("still tx data queued after %u ms\n", timeout); in wait_for_ack() 365 unsigned int queued; in process_one_client() local 367 ret = ioctl(fd, FIONREAD, &queued); in process_one_client() 370 if (queued > expect_le in process_one_client() [all...] |
| /linux/security/integrity/ima/ |
| H A D | ima_asymmetric_keys.c | 33 bool queued = false; in ima_post_key_create_or_update() local 43 queued = ima_queue_key(keyring, payload, payload_len); in ima_post_key_create_or_update() 45 if (queued) in ima_post_key_create_or_update()
|
| H A D | ima_queue_keys.c | 107 bool queued = false; in ima_queue_key() local 117 queued = true; in ima_queue_key() 121 if (!queued) in ima_queue_key() 124 return queued; in ima_queue_key()
|
| /linux/Documentation/userspace-api/media/mediactl/ |
| H A D | media-request-ioc-queue.rst | 34 If the request was successfully queued, then the file descriptor can be 37 If the request was already queued before, then ``EBUSY`` is returned. 42 Once a request is queued, then the driver is required to gracefully handle 49 queued directly and you next try to queue a request, or vice versa. 62 The request was already queued or the application queued the first
|
| H A D | media-request-ioc-reinit.rst | 40 A request can only be re-initialized if it either has not been queued 41 yet, or if it was queued and completed. Otherwise it will set ``errno`` 51 The request is queued but not yet completed.
|
| H A D | request-api.rst | 24 buffer queues since in practice only one buffer would be queued at a time. 59 instead of being immediately applied, and buffers queued to a request do not 60 enter the regular buffer queue until the request itself is queued. 66 queued by calling :ref:`MEDIA_REQUEST_IOC_QUEUE` on the request file descriptor. 68 A queued request cannot be modified anymore. 86 a buffer was queued via a request or vice versa will result in an ``EBUSY`` 109 request that has been queued but not yet completed will return ``EBUSY`` 121 longer in use by the kernel. That is, if the request is queued and then the 165 Once the request is fully prepared, it can be queued to the driver: 245 Once the request is fully prepared, it can be queued to the driver:
|
| /linux/virt/kvm/ |
| H A D | async_pf.c | 111 * need to be flushed (but sanity check that the work wasn't queued). in kvm_flush_and_free_async_pf_work() 150 vcpu->async_pf.queued = 0; in kvm_clear_async_pf_completion_queue() 170 vcpu->async_pf.queued--; in kvm_check_async_pf_completion() 184 if (vcpu->async_pf.queued >= ASYNC_PF_PER_VCPU) in kvm_setup_async_pf() 208 vcpu->async_pf.queued++; in kvm_setup_async_pf() 239 vcpu->async_pf.queued++; in kvm_async_pf_wakeup_all()
|
| /linux/fs/xfs/ |
| H A D | xfs_mru_cache.c | 102 unsigned int queued; /* work has been queued */ member 204 if (!mru->queued) { in _xfs_mru_cache_list_insert() 205 mru->queued = 1; in _xfs_mru_cache_list_insert() 279 mru->queued = next; in _xfs_mru_cache_reap() 280 if ((mru->queued > 0)) { in _xfs_mru_cache_reap() 382 if (mru->queued) { in xfs_mru_cache_flush()
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | null.c | 121 int queued; in ublk_null_queue_io() local 124 queued = null_queue_auto_zc_io(t, q, tag); in ublk_null_queue_io() 126 queued = null_queue_zc_io(t, q, tag); in ublk_null_queue_io() 131 ublk_queued_tgt_io(t, q, tag, queued); in ublk_null_queue_io()
|
| H A D | kublk.h | 376 unsigned tag, int queued) in ublk_queued_tgt_io() 378 if (queued < 0) in ublk_queued_tgt_io() 379 ublk_complete_io(t, q, tag, queued); in ublk_queued_tgt_io() 383 t->io_inflight += queued; in ublk_queued_tgt_io() 384 io->tgt_ios = queued; in ublk_queued_tgt_io() 373 ublk_queued_tgt_io(struct ublk_thread * t,struct ublk_queue * q,unsigned tag,int queued) ublk_queued_tgt_io() argument
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | vidioc-streamon.rst | 51 If ``VIDIOC_STREAMON`` fails then any already queued buffers will remain 52 queued. 63 If buffers have been queued with :ref:`VIDIOC_QBUF` and 65 ``VIDIOC_STREAMON``, then those queued buffers will also be removed from 77 but ``VIDIOC_STREAMOFF`` will return queued buffers to their starting
|
| /linux/Documentation/features/locking/queued-rwlocks/ |
| H A D | arch-support.txt | 2 # Feature name: queued-rwlocks 4 # description: arch supports queued rwlocks
|
| /linux/Documentation/features/locking/queued-spinlocks/ |
| H A D | arch-support.txt | 2 # Feature name: queued-spinlocks 4 # description: arch supports queued spinlocks
|
| /linux/Documentation/usb/ |
| H A D | ohci.rst | 22 - interrupt transfers can be larger, and can be queued 28 types can be queued. That was also true in "usb-ohci", except for interrupt 30 to overhead in IRQ processing. When interrupt transfers are queued, those
|
| /linux/fs/btrfs/ |
| H A D | discard.c | 130 bool queued; in add_to_discard_unused_list() local 134 queued = !list_empty(&block_group->discard_list); in add_to_discard_unused_list() 147 if (!queued) in add_to_discard_unused_list() 159 bool queued = false; in remove_from_discard_list() local 169 queued = !list_empty(&block_group->discard_list); in remove_from_discard_list() 171 if (queued) in remove_from_discard_list()
|
| /linux/drivers/dma/qcom/ |
| H A D | hidma.h | 30 bool queued; /* flag whether this is pending */ member 102 struct list_head queued; member
|
| /linux/Documentation/driver-api/media/drivers/ |
| H A D | pxa_camera.rst | 22 When a buffer is queued, start_streaming is called and the QCI starts. 25 More buffers can be queued while the QCI is started without halting the 91 - first buffer queued for capture 92 Once a first buffer is queued for capture, the QCI is started, but data
|