Home
last modified time | relevance | path

Searched refs:wr_id (Results 1 – 25 of 40) sorted by relevance

12

/linux/drivers/infiniband/sw/rdmavt/
H A Dtrace_tx.h53 __field(u64, wr_id)
74 __entry->wr_id = wqe->wr.wr_id;
95 __entry->wr_id,
121 __field(u64, wr_id)
133 __entry->wr_id = wqe->wr.wr_id;
149 __entry->wr_id,
H A Dtrace_cq.h74 __field(u64, wr_id)
85 __entry->wr_id = wc->wr_id;
98 __entry->wr_id,
H A Dsrq.c201 p->wr_id = wqe->wr_id; in rvt_modify_srq()
/linux/drivers/infiniband/ulp/ipoib/
H A Dipoib_ib.c107 priv->rx_wr.wr_id = id | IPOIB_OP_RECV; in ipoib_ib_post_receive()
177 unsigned int wr_id = wc->wr_id & ~IPOIB_OP_RECV; in ipoib_ib_handle_rx_wc() local
184 wr_id, wc->status); in ipoib_ib_handle_rx_wc()
186 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_ib_handle_rx_wc()
188 wr_id, ipoib_recvq_size); in ipoib_ib_handle_rx_wc()
192 skb = priv->rx_ring[wr_id].skb; in ipoib_ib_handle_rx_wc()
198 wc->status, wr_id, wc->vendor_err); in ipoib_ib_handle_rx_wc()
199 ipoib_ud_dma_unmap_rx(priv, priv->rx_ring[wr_id].mapping); in ipoib_ib_handle_rx_wc()
201 priv->rx_ring[wr_id].skb = NULL; in ipoib_ib_handle_rx_wc()
205 memcpy(mapping, priv->rx_ring[wr_id].mapping, in ipoib_ib_handle_rx_wc()
[all …]
H A Dipoib_cm.c99 priv->cm.rx_wr.wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_srq()
124 wr->wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_nonsrq()
226 ipoib_cm_rx_drain_wr.wr_id = IPOIB_CM_RX_DRAIN_WRID; in ipoib_cm_start_rx_drain()
564 unsigned int wr_id = wc->wr_id & ~(IPOIB_OP_CM | IPOIB_OP_RECV); in ipoib_cm_handle_rx_wc() local
574 wr_id, wc->status); in ipoib_cm_handle_rx_wc()
576 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_cm_handle_rx_wc()
577 if (wr_id == (IPOIB_CM_RX_DRAIN_WRID & ~(IPOIB_OP_CM | IPOIB_OP_RECV))) { in ipoib_cm_handle_rx_wc()
585 wr_id, ipoib_recvq_size); in ipoib_cm_handle_rx_wc()
594 skb = rx_ring[wr_id].skb; in ipoib_cm_handle_rx_wc()
599 wc->status, wr_id, wc->vendor_err); in ipoib_cm_handle_rx_wc()
[all …]
/linux/net/rds/
H A Dib_ring.c156 u32 rds_ib_ring_completed(struct rds_ib_work_ring *ring, u32 wr_id, u32 oldest) in rds_ib_ring_completed() argument
160 if (oldest <= (unsigned long long)wr_id) in rds_ib_ring_completed()
161 ret = (unsigned long long)wr_id - oldest + 1; in rds_ib_ring_completed()
163 ret = ring->w_nr - oldest + (unsigned long long)wr_id + 1; in rds_ib_ring_completed()
166 wr_id, oldest); in rds_ib_ring_completed()
H A Dib_frmr.c157 reg_wr.wr.wr_id = (unsigned long)(void *)ibmr; in rds_ib_post_reg_frmr()
292 s_wr->wr_id = (unsigned long)(void *)ibmr; in rds_ib_post_inv()
328 struct rds_ib_mr *ibmr = (void *)(unsigned long)wc->wr_id; in rds_ib_mr_cqe_handler()
H A Dib_send.c200 send->s_wr.wr_id = i; in rds_ib_send_init_ring()
254 rdsdebug("wc wr_id 0x%llx status %u (%s) byte_len %u imm_data %u\n", in rds_ib_send_cqe_handler()
255 (unsigned long long)wc->wr_id, wc->status, in rds_ib_send_cqe_handler()
260 if (wc->wr_id == RDS_IB_ACK_WR_ID) { in rds_ib_send_cqe_handler()
269 completed = rds_ib_ring_completed(&ic->i_send_ring, wc->wr_id, oldest); in rds_ib_send_cqe_handler()
/linux/net/smc/
H A Dsmc_wr.h92 u64 wr_id, temp_wr_id; in smc_wr_rx_post()
95 wr_id = ++link->wr_rx_id; /* tasklet context, thus not atomic */ in smc_wr_rx_post()
96 temp_wr_id = wr_id; in smc_wr_rx_post()
98 link->wr_rx_ibs[index].wr_id = wr_id; in smc_wr_rx_post()
94 u64 wr_id, temp_wr_id; smc_wr_rx_post() local
H A Dsmc_wr.c41 u64 wr_id; /* work request id sent */ member
66 static inline int smc_wr_tx_find_pending_index(struct smc_link *link, u64 wr_id) in smc_wr_tx_find_pending_index() argument
71 if (link->wr_tx_pends[i].wr_id == wr_id) in smc_wr_tx_find_pending_index()
94 pnd_snd_idx = smc_wr_tx_find_pending_index(link, wc->wr_id); in smc_wr_tx_process_cqe()
97 link->wr_tx_v2_pend->wr_id != wc->wr_id) in smc_wr_tx_process_cqe()
205 u64 wr_id; in smc_wr_tx_get_free_slot() local
229 wr_id = smc_wr_tx_get_next_wr_id(link); in smc_wr_tx_get_free_slot()
231 wr_pend->wr_id = wr_id; in smc_wr_tx_get_free_slot()
236 wr_ib->wr_id = wr_id; in smc_wr_tx_get_free_slot()
251 u64 wr_id; in smc_wr_tx_get_v2_slot() local
[all …]
/linux/drivers/infiniband/hw/mlx5/
H A Dgsi.c74 u64 wr_id; in handle_single_completion() local
79 wr_id = wr->wc.wr_id; in handle_single_completion()
81 wr->wc.wr_id = wr_id; in handle_single_completion()
380 gsi_wr->wc.wr_id = wr->wr.wr_id; in mlx5_ib_add_outstanding_wr()
396 { .wr_id = wr->wr.wr_id }, in mlx5_ib_gsi_silent_drop()
H A Dwr.c760 u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) in mlx5r_finish_wqe() argument
771 qp->sq.wrid[idx] = wr_id; in mlx5r_finish_wqe()
836 mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx, wr->wr_id, in handle_psv()
878 wr->wr_id, nreq, fence, MLX5_OPCODE_UMR); in handle_reg_mr_integrity()
909 mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx, wr->wr_id, in handle_reg_mr_integrity()
1185 mlx5r_finish_wqe(qp, ctrl, seg, size, cur_edge, idx, wr->wr_id, in mlx5_ib_post_send()
1264 qp->rq.wrid[ind] = wr->wr_id; in mlx5_ib_post_recv()
H A Dwr.h101 u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode);
/linux/drivers/infiniband/hw/cxgb4/
H A Dcq.c469 srq->sw_rq[srq->pidx].wr_id = pwr->wr_id; in post_pending_srq_wrs()
476 (unsigned long long)pwr->wr_id); in post_pending_srq_wrs()
494 u64 wr_id; in reap_srq_cqe() local
497 wr_id = srq->sw_rq[rel_idx].wr_id; in reap_srq_cqe()
503 (unsigned long long)srq->sw_rq[rel_idx].wr_id); in reap_srq_cqe()
511 srq->sw_rq[srq->cidx].wr_id); in reap_srq_cqe()
522 (unsigned long long)srq->sw_rq[rel_idx].wr_id); in reap_srq_cqe()
525 return wr_id; in reap_srq_cqe()
717 *cookie = wq->sq.sw_sq[wq->sq.cidx].wr_id; in poll_cq()
724 *cookie = wq->rq.sw_rq[wq->rq.cidx].wr_id; in poll_cq()
[all …]
H A Dqp.c714 swsqe->wr_id = wr->wr_id; in post_write_cmpl()
738 swsqe->wr_id = wr->next->wr_id; in post_write_cmpl()
992 cqe.u.drain_cookie = wr->wr_id; in complete_sq_drain_wr()
1043 cqe.u.drain_cookie = wr->wr_id; in complete_rq_drain_wr()
1231 swsqe->wr_id = wr->wr_id; in c4iw_post_send()
1241 (unsigned long long)wr->wr_id, qhp->wq.sq.pidx, in c4iw_post_send()
1305 qhp->wq.rq.sw_rq[qhp->wq.rq.pidx].wr_id = wr->wr_id; in c4iw_post_receive()
1322 (unsigned long long)wr->wr_id, qhp->wq.rq.pidx); in c4iw_post_receive()
1339 u64 wr_id, u8 len16) in defer_srq_wr() argument
1346 (unsigned long long)wr_id, srq->pending_cidx, in defer_srq_wr()
[all …]
/linux/include/uapi/rdma/
H A Dvmw_pvrdma-abi.h230 __aligned_u64 wr_id; /* wr id */ member
238 __aligned_u64 wr_id; /* wr id */ member
291 __aligned_u64 wr_id; member
H A Drdma_user_rxe.h76 __aligned_u64 wr_id; member
176 __aligned_u64 wr_id; member
H A Drvt-abi.h44 __u64 wr_id; member
/linux/drivers/infiniband/hw/qedr/
H A Dqedr_roce_cm.c582 qp->wqe_wr_id[qp->sq.prod].wr_id = wr->wr_id; in qedr_gsi_post_send()
586 wr->wr_id); in qedr_gsi_post_send()
652 qp->rqe_wr_id[qp->rq.prod].wr_id = wr->wr_id; in qedr_gsi_post_recv()
683 wc[i].wr_id = qp->rqe_wr_id[qp->rq.cons].wr_id; in qedr_gsi_poll_cq()
711 wc[i].wr_id = qp->wqe_wr_id[qp->sq.cons].wr_id; in qedr_gsi_poll_cq()
/linux/drivers/infiniband/hw/mlx4/
H A Dmad.c663 wr.wr.wr_id = ((u64) tun_tx_ix) | MLX4_TUN_SET_WRID_QPN(dest_qpt); in mlx4_ib_send_to_slave()
1333 recv_wr.wr_id = (u64) index | MLX4_TUN_WRID_RECV | in mlx4_ib_post_pv_qp_buf()
1441 wr.wr.wr_id = ((u64) wire_tx_ix) | MLX4_TUN_SET_WRID_QPN(src_qpnum); in mlx4_ib_send_to_wire()
1480 struct mlx4_ib_demux_pv_qp *tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc->wr_id)]; in mlx4_ib_multiplex_mad()
1481 int wr_ix = wc->wr_id & (MLX4_NUM_TUNNEL_BUFS - 1); in mlx4_ib_multiplex_mad()
1742 tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc.wr_id)]; in mlx4_ib_tunnel_comp_worker()
1748 wc.wr_id & in mlx4_ib_tunnel_comp_worker()
1752 "buf:%lld\n", wc.wr_id); in mlx4_ib_tunnel_comp_worker()
1755 rdma_destroy_ah(tun_qp->tx_ring[wc.wr_id & in mlx4_ib_tunnel_comp_worker()
1757 tun_qp->tx_ring[wc.wr_id & (MLX4_NUM_TUNNEL_BUFS - 1)].ah in mlx4_ib_tunnel_comp_worker()
[all …]
/linux/include/rdma/
H A Drdmavt_qp.h837 u64 wr_id; in rvt_qp_complete_swqe() local
850 wr_id = wqe->wr.wr_id; in rvt_qp_complete_swqe()
859 .wr_id = wr_id, in rvt_qp_complete_swqe()
/linux/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.h87 u64 wr_id; member
126 u64 wr_id; /* work request ID */ member
149 u64 wr_id; /* work request ID from WQE */ member
H A Dehea_main.c490 rwqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, wqe_type) in ehea_refill_rq_def()
567 int skb_index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in get_skb_by_index()
781 swqe->wr_id = SWQE_RESTART_CHECK; in check_sqs()
820 if (cqe->wr_id == SWQE_RESTART_CHECK) { in ehea_proc_cqes()
843 if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id) in ehea_proc_cqes()
846 index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in ehea_proc_cqes()
852 swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id); in ehea_proc_cqes()
2043 swqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, EHEA_SWQE3_TYPE) in ehea_start_xmit()
2046 swqe->wr_id |= EHEA_BMASK_SET(EHEA_WR_ID_REFILL, in ehea_start_xmit()
2053 swqe->wr_id = in ehea_start_xmit()
[all …]
/linux/drivers/infiniband/hw/irdma/
H A Duk.c168 qp->sq_wrtrk_array[*wqe_idx].wrid = info->wr_id; in irdma_qp_get_next_send_wqe()
456 set_64bit_val(wqe, 16, (u64)info->wr_id); in irdma_uk_srq_post_receive()
1005 qp->rq_wrid_array[wqe_idx] = info->wr_id; in irdma_uk_post_receive()
1288 get_64bit_val(cqe, 8, &info->wr_id); in irdma_uk_cq_poll_cmpl()
1315 info->wr_id = qp->rq_wrid_array[qp->rq_ring.tail]; in irdma_uk_cq_poll_cmpl()
1318 info->wr_id = qp->rq_wrid_array[array_idx]; in irdma_uk_cq_poll_cmpl()
1354 info->wr_id = qp->sq_wrtrk_array[wqe_idx].wrid; in irdma_uk_cq_poll_cmpl()
1384 info->wr_id = qp->sq_wrtrk_array[tail].wrid; in irdma_uk_cq_poll_cmpl()
1816 int irdma_nop(struct irdma_qp_uk *qp, u64 wr_id, bool signaled, bool post_sq) in irdma_nop() argument
1823 info.wr_id = wr_id; in irdma_nop()
/linux/drivers/infiniband/sw/siw/
H A Dsiw_cq.c58 wc->wr_id = cqe->id; in siw_reap_cqe()

12