Searched refs:sq_ring (Results 1 – 8 of 8) sorted by relevance
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | uk.c | 65 if (!qp->sq_ring.head) in irdma_nop_1() 68 wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring); in irdma_nop_1() 100 wqe_idx = (qp_wqe_idx + 128) % qp->sq_ring.size; in irdma_clr_wqes() 139 (IRDMA_RING_CURRENT_HEAD(qp->sq_ring) % in irdma_qp_get_next_send_wqe() 143 if (quanta > IRDMA_SQ_RING_FREE_QUANTA(qp->sq_ring)) in irdma_qp_get_next_send_wqe() 148 IRDMA_SQ_RING_FREE_QUANTA(qp->sq_ring)) in irdma_qp_get_next_send_wqe() 153 IRDMA_RING_MOVE_HEAD_NOCHECK(qp->sq_ring); in irdma_qp_get_next_send_wqe() 157 *wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring); in irdma_qp_get_next_send_wqe() 161 IRDMA_RING_MOVE_HEAD_BY_COUNT_NOCHECK(qp->sq_ring, quanta); in irdma_qp_get_next_send_wqe() 165 (IRDMA_RING_CURRENT_HEAD(qp->sq_ring) & 1)) { in irdma_qp_get_next_send_wqe() [all …]
|
| H A D | utils.c | 557 pending_work = IRDMA_RING_USED_QUANTA(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op() 558 wqe_idx = IRDMA_RING_CURRENT_TAIL(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op() 564 wqe_idx = (wqe_idx + 1) % IRDMA_RING_SIZE(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op() 2416 struct irdma_ring *sq_ring = &qp->sq_ring; in irdma_generate_flush_completions() local 2432 while (IRDMA_RING_MORE_WORK(*sq_ring)) { in irdma_generate_flush_completions() 2440 wqe_idx = sq_ring->tail; in irdma_generate_flush_completions() 2449 IRDMA_RING_SET_TAIL(*sq_ring, in irdma_generate_flush_completions() 2450 sq_ring->tail + qp->sq_wrtrk_array[sq_ring->tail].quanta); in irdma_generate_flush_completions()
|
| H A D | ctrl.c | 420 qp->hw_sq_size = irdma_get_encoded_wqe_size(qp->qp_uk.sq_ring.size, in irdma_sc_qp_init() 424 qp->hw_sq_size, qp->qp_uk.sq_ring.size); in irdma_sc_qp_init() 2497 cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size); in irdma_sc_gather_stats() 3291 IRDMA_RING_MOVE_TAIL(cqp->sq_ring); in irdma_cqp_poll_registers() 3653 IRDMA_RING_INIT(cqp->sq_ring, cqp->sq_size); in irdma_sc_cqp_init() 3798 writel(IRDMA_RING_CURRENT_HEAD(cqp->sq_ring), cqp->dev->cqp_db); in irdma_sc_cqp_post_sq() 3802 cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size); in irdma_sc_cqp_post_sq() 3818 if (IRDMA_RING_FULL_ERR(cqp->sq_ring)) { in irdma_sc_cqp_get_next_send_wqe_idx() 3821 cqp->sq_ring.head, cqp->sq_ring.tail, in irdma_sc_cqp_get_next_send_wqe_idx() 3822 cqp->sq_ring.size); in irdma_sc_cqp_get_next_send_wqe_idx() [all …]
|
| H A D | puda.c | 196 *wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring); in irdma_puda_get_next_send_wqe() 199 IRDMA_RING_MOVE_HEAD(qp->sq_ring, ret_code); in irdma_puda_get_next_send_wqe() 399 IRDMA_RING_SET_TAIL(qp->sq_ring, info.wqe_idx); in irdma_puda_poll_cmpl() 687 IRDMA_RING_INIT(ukqp->sq_ring, ukqp->sq_size); in irdma_puda_qp_create()
|
| H A D | user.h | 502 struct irdma_ring sq_ring; member
|
| H A D | hw.c | 2627 if (IRDMA_RING_MORE_WORK(qp->qp_uk.sq_ring)) { in irdma_hw_flush_wqes_callback() 2690 if (IRDMA_RING_MORE_WORK(qp->qp_uk.sq_ring)) { in irdma_hw_flush_wqes() 2728 if (!IRDMA_RING_MORE_WORK(qp->qp_uk.sq_ring)) in irdma_hw_flush_wqes()
|
| H A D | type.h | 430 struct irdma_ring sq_ring; member
|
| /linux/tools/testing/selftests/x86/ |
| H A D | lam.c | 113 struct io_uring_queue sq_ring; member 486 struct io_uring_queue *sring = &s->sq_ring; in mmap_io_uring() 529 s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in mmap_io_uring() 532 if (s->sq_ring.queue.sqes == MAP_FAILED) { in mmap_io_uring() 608 struct io_uring_queue *sring = &ring->sq_ring; in handle_uring_sq() 638 index = tail & *ring->sq_ring.ring_mask; in handle_uring_sq() 640 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq()
|