| /linux/tools/virtio/ringtest/ |
| H A D | main.c | 109 int completed = 0; in run_guest() local 121 completed_before = completed; in run_guest() 124 started - completed < max_outstanding) { in run_guest() 140 ++completed; in run_guest() 141 if (__builtin_expect(completed == bufs, false)) in run_guest() 146 if (completed == completed_before) in run_guest() 148 assert(completed <= bufs); in run_guest() 168 int completed = 0; in run_host() local 183 completed_before = completed; in run_host() 187 ++completed; in run_host() [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | opfn.c | 59 priv->opfn.requested == priv->opfn.completed || priv->opfn.curr) in opfn_conn_request() 62 mask = priv->opfn.requested & ~priv->opfn.completed; in opfn_conn_request() 65 priv->opfn.completed |= OPFN_CODE(capcode); in opfn_conn_request() 76 priv->opfn.completed |= OPFN_CODE(capcode); in opfn_conn_request() 158 if (priv->opfn.completed & OPFN_CODE(capcode)) { in opfn_conn_response() 163 priv->opfn.completed &= ~OPFN_CODE(capcode); in opfn_conn_response() 169 priv->opfn.completed |= OPFN_CODE(capcode); in opfn_conn_response() 202 priv->opfn.completed |= OPFN_CODE(capcode); in opfn_conn_reply() 229 while (priv->opfn.completed) { in opfn_conn_error() 230 capcode = priv->opfn.completed & ~(priv->opfn.completed - 1); in opfn_conn_error() [all …]
|
| /linux/drivers/hv/ |
| H A D | mshv_common.c | 35 u16 completed = 0; in hv_call_get_vp_registers() local 62 completed = hv_repcomp(status); in hv_call_get_vp_registers() 63 for (i = 0; i < completed; ++i) in hv_call_get_vp_registers() 66 registers += completed; in hv_call_get_vp_registers() 67 remaining -= completed; in hv_call_get_vp_registers() 80 u16 completed = 0; in hv_call_set_vp_registers() local 105 completed = hv_repcomp(status); in hv_call_set_vp_registers() 106 registers += completed; in hv_call_set_vp_registers() 107 remaining -= completed; in hv_call_set_vp_registers()
|
| H A D | mshv_root_hv_call.c | 47 u16 completed; in hv_call_withdraw_memory() local 71 completed = hv_repcomp(status); in hv_call_withdraw_memory() 73 for (i = 0; i < completed; i++) in hv_call_withdraw_memory() 82 remaining -= completed; in hv_call_withdraw_memory() 207 ulong i, completed, remain = page_count - done; in hv_do_map_gpa_hcall() local 239 completed = hv_repcomp(status); in hv_do_map_gpa_hcall() 252 done += completed; in hv_do_map_gpa_hcall() 309 ulong completed, remain = page_count - done; in hv_call_unmap_gpa_pages() local 322 completed = hv_repcomp(status); in hv_call_unmap_gpa_pages() 328 done += completed; in hv_call_unmap_gpa_pages() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | cq.c | 95 int i, n, completed = 0; in __ib_process_cq() local 105 budget - completed), wcs)) > 0) { in __ib_process_cq() 115 completed += n; in __ib_process_cq() 117 if (n != batch || (budget != -1 && completed >= budget)) in __ib_process_cq() 121 return completed; in __ib_process_cq() 155 int completed; in ib_poll_handler() local 157 completed = __ib_process_cq(cq, budget, cq->wc, IB_POLL_BATCH); in ib_poll_handler() 158 if (completed < budget) { in ib_poll_handler() 167 rdma_dim(dim, completed); in ib_poll_handler() 169 return completed; in ib_poll_handler() [all …]
|
| /linux/lib/ |
| H A D | dynamic_queue_limits.c | 86 unsigned int ovlimit, completed, num_queued; in dql_completed() local 101 completed = dql->num_completed + count; in dql_completed() 104 inprogress = num_queued - completed; in dql_completed() 106 all_prev_completed = AFTER_EQ(completed, dql->prev_num_queued); in dql_completed() 125 limit += POSDIFF(completed, dql->prev_num_queued) + in dql_completed() 155 2 * (completed - dql->num_completed)); in dql_completed() 180 dql->adj_limit = limit + completed; in dql_completed() 183 dql->num_completed = completed; in dql_completed()
|
| /linux/tools/virtio/ |
| H A D | vhost_net_test.c | 37 long completed; member 285 long completed_before = vq->completed; in run_tx_test() 290 (vq->started - vq->completed) < 1) { in run_tx_test() 317 ++vq->completed; in run_tx_test() 322 if (vq->completed == completed_before && vq->started == started_before) in run_tx_test() 325 assert(vq->completed <= bufs); in run_tx_test() 327 if (vq->completed == bufs) in run_tx_test() 339 spurious, vq->started, vq->completed); in run_tx_test() 352 long completed_before = vq->completed; in run_rx_test() 356 (vq->started - vq->completed) < 1) { in run_rx_test() [all …]
|
| H A D | virtio_test.c | 173 long started = 0, completed = 0, next_reset = reset_n; in run_test() local 188 completed_before = completed; in run_test() 191 const bool reset = completed > next_reset; in run_test() 196 (started - completed) < batch) { in run_test() 229 ++completed; in run_test() 251 started = completed; in run_test() 252 while (completed > next_reset) in run_test() 253 next_reset += completed; in run_test() 256 if (completed == completed_before && started == started_before) in run_test() 258 assert(completed <= bufs); in run_test() [all …]
|
| /linux/drivers/media/mc/ |
| H A D | mc-request.c | 388 obj->completed = false; in media_request_object_init() 434 bool completed = false; in media_request_object_unbind() local 450 if (!obj->completed) in media_request_object_unbind() 462 completed = true; in media_request_object_unbind() 470 if (completed) in media_request_object_unbind() 479 bool completed = false; in media_request_object_complete() local 482 if (obj->completed) in media_request_object_complete() 484 obj->completed = true; in media_request_object_complete() 492 completed = true; in media_request_object_complete() 496 if (completed) in media_request_object_complete() [all …]
|
| /linux/Documentation/ABI/testing/ |
| H A D | procfs-diskstats | 13 4 reads completed successfully 17 8 writes completed 30 15 discards completed successfully 39 19 flush requests completed successfully
|
| H A D | sysfs-bus-fsi-devices-sbefifo | 8 occurred and no transfers have completed since the timeout. A 10 has, more recent transfers have completed successfully.
|
| /linux/drivers/i2c/algos/ |
| H A D | i2c-algo-pca.c | 192 int completed = 1; in pca_xfer() local 238 completed = pca_start(adap); in pca_xfer() 243 completed = pca_address(adap, msg); in pca_xfer() 249 completed = pca_tx_byte(adap, in pca_xfer() 258 completed = pca_repeated_start(adap); in pca_xfer() 268 completed = pca_rx_ack(adap, msg->len > 1); in pca_xfer() 275 completed = pca_rx_ack(adap, in pca_xfer() 283 completed = pca_repeated_start(adap); in pca_xfer() 318 completed = pca_repeated_start(adap); in pca_xfer() 345 if (!completed) in pca_xfer()
|
| /linux/Documentation/userspace-api/media/mediactl/ |
| H A D | media-request-ioc-reinit.rst | 36 This avoids having to :c:func:`close()` a completed 37 request and allocate a new request. Instead the completed request can just 41 yet, or if it was queued and completed. Otherwise it will set ``errno`` 51 The request is queued but not yet completed.
|
| /linux/drivers/s390/scsi/ |
| H A D | zfcp_qdio.c | 79 int completed; in zfcp_qdio_request_tasklet() local 81 completed = qdio_inspect_output_queue(cdev, 0, &start, &error); in zfcp_qdio_request_tasklet() 82 if (completed > 0) { in zfcp_qdio_request_tasklet() 87 zfcp_qdio_zero_sbals(qdio->req_q, start, completed); in zfcp_qdio_request_tasklet() 92 atomic_add(completed, &qdio->req_q_free); in zfcp_qdio_request_tasklet() 166 int completed; in zfcp_qdio_irq_tasklet() local 172 completed = qdio_inspect_input_queue(cdev, 0, &start, &error); in zfcp_qdio_irq_tasklet() 173 if (completed < 0) in zfcp_qdio_irq_tasklet() 175 if (completed > 0) in zfcp_qdio_irq_tasklet() 176 zfcp_qdio_int_resp(cdev, error, 0, start, completed, in zfcp_qdio_irq_tasklet()
|
| /linux/drivers/vdpa/mlx5/core/ |
| H A D | resources.c | 334 int *completed) in issue_async_cmd() argument 347 if (*completed < issued) { in issue_async_cmd() 349 wait_for_completion(&cmds[*completed].cmd_done); in issue_async_cmd() 350 (*completed)++; in issue_async_cmd() 359 (*completed)++; in issue_async_cmd() 370 int completed = 0; in mlx5_vdpa_exec_async_cmds() local 379 err = issue_async_cmd(mvdev, cmds, issued, &completed); in mlx5_vdpa_exec_async_cmds() 389 while (completed < issued) in mlx5_vdpa_exec_async_cmds() 390 wait_for_completion(&cmds[completed++].cmd_done); in mlx5_vdpa_exec_async_cmds()
|
| /linux/drivers/scsi/aacraid/ |
| H A D | dpcsup.c | 363 int completed = 0; in aac_intr_normal() local 369 completed = 1; in aac_intr_normal() 382 if (completed) in aac_intr_normal() 416 int completed = 0; in aac_intr_normal() local 422 completed = 1; in aac_intr_normal() 435 if (completed) in aac_intr_normal()
|
| /linux/drivers/usb/gadget/udc/ |
| H A D | pxa27x_udc.c | 871 int count, is_short, completed = 0; in read_fifo() local 884 completed = 1; in read_fifo() 889 return completed; in read_fifo() 907 int count, is_short, is_last = 0, completed = 0, totcount = 0; in write_fifo() local 947 completed = 1; in write_fifo() 956 return completed; in write_fifo() 972 int count, is_short, completed = 0; in read_ep0_fifo() local 985 completed = 1; in read_ep0_fifo() 990 return completed; in read_ep0_fifo() 1936 int completed = 0; in handle_ep0() local [all …]
|
| /linux/Documentation/driver-api/dmaengine/ |
| H A D | pxa_dma.rst | 110 Each time a transfer is completed on a channel, an interrupt might be 113 transfer being completed into the physical channel's completion mark. 117 any lock to find out what is the latest completed transfer in a running 132 If a transfer is completed while this handling is done, a dma irq will 137 Residue granularity will be descriptor based. The issued but not completed 183 - completed queue : empty 187 It should be noted that after tx3 is completed, the channel is stopped, and
|
| /linux/drivers/block/mtip32xx/ |
| H A D | mtip32xx.h | 174 u32 completed; member 181 mtip_workq_sdbfx(w->port, group, w->completed); \ 334 void __iomem *completed[MTIP_MAX_SLOT_GROUPS]; member
|
| /linux/drivers/iio/adc/ |
| H A D | fsl-imx25-gcq.c | 39 struct completion completed; member 92 complete(&priv->completed); in mx25_gcq_irq() 124 &priv->completed, MX25_GCQ_TIMEOUT); in mx25_gcq_get_raw_value() 321 init_completion(&priv->completed); in mx25_gcq_probe()
|
| /linux/drivers/macintosh/ |
| H A D | adb.c | 611 struct adb_request *completed; member 635 struct adb_request **ap = &state->completed; in adb_write_done() 685 state->completed = NULL; in adb_open() 704 && state->completed == NULL) { in adb_release() 736 req = state->completed; in adb_read() 738 state->completed = req->next; in adb_read()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | gsi.c | 38 bool completed:1; member 58 if (!wr->completed) in generate_completions() 62 wr->completed = false; in generate_completions() 78 wr->completed = true; in handle_single_completion() 383 gsi_wr->completed = true; in mlx5_ib_add_outstanding_wr()
|
| /linux/arch/um/drivers/ |
| H A D | virtio_pcidev.c | 180 void *completed = virtqueue_get_buf(dev->cmd_vq, &len); in virtio_pcidev_send_cmd() local 182 if (completed == buf) in virtio_pcidev_send_cmd() 185 if (completed) in virtio_pcidev_send_cmd() 186 virtio_pcidev_free_buf(dev, completed); in virtio_pcidev_send_cmd()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | send.c | 640 struct mlx5hws_completed_poll *comp = &queue->completed; in hws_send_engine_poll_list() 1022 kfree(queue->completed.entries); in mlx5hws_send_queue_close() 1036 queue->completed.entries = kzalloc_objs(queue->completed.entries[0], in mlx5hws_send_queue_open() 1038 if (!queue->completed.entries) in mlx5hws_send_queue_open() 1041 queue->completed.pi = 0; in mlx5hws_send_queue_open() 1042 queue->completed.ci = 0; in mlx5hws_send_queue_open() 1043 queue->completed.mask = queue->num_entries - 1; in mlx5hws_send_queue_open() 1051 kfree(queue->completed.entries); in mlx5hws_send_queue_open()
|
| /linux/Documentation/scsi/ |
| H A D | scsi_eh.rst | 15 [1-2] How do scmd's get completed? 45 1.2 How do scmd's get completed? 71 the completed request by calling blk_end_request and 104 eh_timed_out() completed the command. 153 completed a scmd with error status, the LLDD and lower layers are 318 This action is taken for each error-completed 363 as we explicitly choose error-completed scmds, it is known 387 to choose error-completed scmds.
|