| /linux/drivers/infiniband/ulp/isert/ |
| H A D | ib_isert.c | 781 struct ib_send_wr send_wr; in isert_login_post_send() local 789 send_wr.next = NULL; in isert_login_post_send() 790 send_wr.wr_cqe = &tx_desc->tx_cqe; in isert_login_post_send() 791 send_wr.sg_list = tx_desc->tx_sg; in isert_login_post_send() 792 send_wr.num_sge = tx_desc->num_sge; in isert_login_post_send() 793 send_wr.opcode = IB_WR_SEND; in isert_login_post_send() 794 send_wr.send_flags = IB_SEND_SIGNALED; in isert_login_post_send() 796 ret = ib_post_send(isert_conn->qp, &send_wr, NULL); in isert_login_post_send() 862 struct ib_send_wr *send_wr) in isert_init_send_wr() argument 867 send_wr->wr_cqe = &tx_desc->tx_cqe; in isert_init_send_wr() [all …]
|
| H A D | ib_isert.h | 132 struct ib_send_wr send_wr; member
|
| /linux/drivers/infiniband/core/ |
| H A D | mad.c | 62 struct ib_ud_wr *wr = &mad_send_wr->send_wr; in create_mad_addr_info() 634 struct ib_ud_wr *send_wr = &mad_send_wr->send_wr; in handle_outgoing_dr_smp() local 643 port_num = send_wr->port_num; in handle_outgoing_dr_smp() 716 send_wr->wr.wr_cqe, drslid, in handle_outgoing_dr_smp() 717 send_wr->pkey_index, in handle_outgoing_dr_smp() 718 send_wr->port_num, &mad_wc); in handle_outgoing_dr_smp() 777 local->mad_send_wr->send_wr.pkey_index = out_mad_pkey_index; in handle_outgoing_dr_smp() 815 static int alloc_send_rmpp_list(struct ib_mad_send_wr_private *send_wr, in alloc_send_rmpp_list() argument 818 struct ib_mad_send_buf *send_buf = &send_wr->send_buf; in alloc_send_rmpp_list() 826 pad = send_wr->pad; in alloc_send_rmpp_list() [all …]
|
| H A D | cma_trace.h | 186 __field(u32, send_wr) 198 __entry->send_wr = qp_init_attr->cap.max_send_wr; 218 rdma_show_qp_type(__entry->qp_type), __entry->send_wr,
|
| H A D | agent.c | 137 mad_send_wr->send_wr.port_num = port_num; in agent_send_response()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | svc_rdma_sendto.c | 352 struct ib_send_wr *send_wr = &ctxt->sc_send_wr; in svc_rdma_post_send() 361 send_wr->sg_list[0].addr, in svc_rdma_post_send() 362 send_wr->sg_list[0].length, in svc_rdma_post_send() 895 struct ib_send_wr *send_wr = &sctxt->sc_send_wr; in svc_rdma_send_reply_msg() 909 send_wr->opcode = IB_WR_SEND_WITH_INV; 910 send_wr->ex.invalidate_rkey = rctxt->rc_inv_rkey; 912 send_wr->opcode = IB_WR_SEND; 342 struct ib_send_wr *send_wr = &ctxt->sc_send_wr; svc_rdma_post_send() local 885 struct ib_send_wr *send_wr = &sctxt->sc_send_wr; svc_rdma_send_reply_msg() local
|
| H A D | frwr_ops.c | 386 struct ib_send_wr *post_wr, *send_wr = &req->rl_wr; in frwr_send() local 393 post_wr = send_wr; in frwr_send() 408 send_wr->send_flags |= IB_SEND_SIGNALED; in frwr_send() 412 send_wr->send_flags &= ~IB_SEND_SIGNALED; in frwr_send()
|
| /linux/drivers/infiniband/ulp/iser/ |
| H A D | iscsi_iser.h | 144 #define ISER_GET_MAX_XMIT_CMDS(send_wr) ((send_wr \ argument 247 struct ib_send_wr send_wr; member
|
| H A D | iser_memory.c | 281 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr() 328 wr->wr.next = &tx_desc->send_wr; in iser_fast_reg_mr()
|
| H A D | iser_verbs.c | 859 struct ib_send_wr *wr = &tx_desc->send_wr; in iser_post_send()
|
| /linux/drivers/nvme/target/ |
| H A D | rdma.c | 63 struct ib_send_wr send_wr; member 420 r->send_wr.wr_cqe = &r->send_cqe; in nvmet_rdma_alloc_rsp() 421 r->send_wr.sg_list = &r->send_sge; in nvmet_rdma_alloc_rsp() 422 r->send_wr.num_sge = 1; in nvmet_rdma_alloc_rsp() 423 r->send_wr.send_flags = IB_SEND_SIGNALED; in nvmet_rdma_alloc_rsp() 716 rsp->send_wr.opcode = IB_WR_SEND_WITH_INV; in nvmet_rdma_queue_response() 717 rsp->send_wr.ex.invalidate_rkey = rsp->invalidate_rkey; in nvmet_rdma_queue_response() 719 rsp->send_wr.opcode = IB_WR_SEND; in nvmet_rdma_queue_response() 728 cm_id->port_num, NULL, &rsp->send_wr); in nvmet_rdma_queue_response() 730 first_wr = &rsp->send_wr; in nvmet_rdma_queue_response() [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | wr.c | 416 static int set_sig_data_segment(const struct ib_send_wr *send_wr, in set_sig_data_segment() argument 557 static int set_pi_umr_wr(const struct ib_send_wr *send_wr, in set_pi_umr_wr() argument 561 const struct ib_reg_wr *wr = reg_wr(send_wr); in set_pi_umr_wr() 569 if (unlikely(send_wr->num_sge != 0) || in set_pi_umr_wr() 599 ret = set_sig_data_segment(send_wr, wr->mr, sig_attrs, qp, seg, size, in set_pi_umr_wr()
|
| /linux/drivers/infiniband/ulp/srpt/ |
| H A D | ib_srpt.c | 2875 struct ib_send_wr send_wr, *first_wr = &send_wr; in srpt_queue_response() local 2938 send_wr.next = NULL; in srpt_queue_response() 2939 send_wr.wr_cqe = &ioctx->ioctx.cqe; in srpt_queue_response() 2940 send_wr.sg_list = &sge; in srpt_queue_response() 2941 send_wr.num_sge = 1; in srpt_queue_response() 2942 send_wr.opcode = IB_WR_SEND; in srpt_queue_response() 2943 send_wr.send_flags = IB_SEND_SIGNALED; in srpt_queue_response()
|
| /linux/include/rdma/ |
| H A D | ib_verbs.h | 2399 int (*post_send)(struct ib_qp *qp, const struct ib_send_wr *send_wr, 3971 * @send_wr: A list of work requests to post on the send queue. in ib_post_recv() 3981 const struct ib_send_wr *send_wr, 3986 return qp->device->ops.post_send(qp, send_wr, bad_send_wr ? : &dummy); in ib_alloc_cq() 3954 ib_post_send(struct ib_qp * qp,const struct ib_send_wr * send_wr,const struct ib_send_wr ** bad_send_wr) ib_post_send() argument
|
| /linux/include/uapi/rdma/ |
| H A D | ib_user_verbs.h | 859 struct ib_uverbs_send_wr send_wr[]; member
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_hw_v2.c | 3689 struct ib_send_wr *send_wr; in free_mr_post_send_lp_wqe() local 3692 send_wr = &rdma_wr.wr; in free_mr_post_send_lp_wqe() 3693 send_wr->opcode = IB_WR_RDMA_WRITE; in free_mr_post_send_lp_wqe() 3695 ret = hns_roce_v2_post_send(&hr_qp->ibqp, send_wr, &bad_wr); in free_mr_post_send_lp_wqe()
|