| /linux/fs/smb/smbdirect/ |
| H A D | rw.c | 120 struct ib_send_wr *first_wr; in smbdirect_connection_rdma_xmit() local 222 first_wr = NULL; in smbdirect_connection_rdma_xmit() 224 first_wr = rdma_rw_ctx_wrs(&msg->rdma_ctx, in smbdirect_connection_rdma_xmit() 228 first_wr); in smbdirect_connection_rdma_xmit() 231 ret = ib_post_send(sc->ib.qp, first_wr, NULL); in smbdirect_connection_rdma_xmit()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | svc_rdma_rw.c | 405 struct ib_send_wr *first_wr; in svc_rdma_post_chunk_ctxt() 416 first_wr = NULL; in svc_rdma_post_chunk_ctxt() 422 first_wr = rdma_rw_ctx_wrs(&ctxt->rw_ctx, rdma->sc_qp, in svc_rdma_post_chunk_ctxt() 423 rdma->sc_port_num, cqe, first_wr); in svc_rdma_post_chunk_ctxt() 432 ret = ib_post_send(rdma->sc_qp, first_wr, &bad_wr); in svc_rdma_post_chunk_ctxt() 435 first_wr, cc->cc_sqecount, in svc_rdma_post_chunk_ctxt() 629 struct ib_send_wr *first_wr; in svc_rdma_send_write_chunk() 633 first_wr = sctxt->sc_wr_chain; in svc_rdma_send_write_chunk() 639 first_wr = rdma_rw_ctx_wrs(&rwc->rw_ctx, rdma->sc_qp, in svc_rdma_send_write_chunk() 640 rdma->sc_port_num, cqe, first_wr); in svc_rdma_send_write_chunk() 386 struct ib_send_wr *first_wr; svc_rdma_post_chunk_ctxt() local 702 struct ib_send_wr *first_wr; svc_rdma_prepare_reply_chunk() local [all...] |
| H A D | svc_rdma_sendto.c | 383 * @first_wr: first WR in the chain in svc_rdma_post_send() 394 const struct ib_send_wr *first_wr, 403 if (bad_wr != first_wr) in svc_rdma_encode_read_list() 459 struct ib_send_wr *first_wr = ctxt->sc_wr_chain; in svc_rdma_encode_write_chunk() 461 const struct ib_send_wr *bad_wr = first_wr; in svc_rdma_encode_write_chunk() 478 ret = ib_post_send(rdma->sc_qp, first_wr, &bad_wr); in svc_rdma_encode_write_chunk() 481 first_wr, sqecount, ret); in svc_rdma_encode_write_chunk() 341 struct ib_send_wr *first_wr = ctxt->sc_wr_chain; svc_rdma_post_send() local
|
| /linux/drivers/infiniband/core/ |
| H A D | rw.c | 872 struct ib_send_wr *first_wr, *last_wr; in rdma_rw_ctx_wrs() local 885 first_wr = &ctx->reg[0].inv_wr; in rdma_rw_ctx_wrs() 887 first_wr = &ctx->reg[0].reg_wr.wr; in rdma_rw_ctx_wrs() 891 first_wr = &ctx->iova.wr.wr; in rdma_rw_ctx_wrs() 895 first_wr = &ctx->map.wrs[0].wr; in rdma_rw_ctx_wrs() 899 first_wr = &ctx->single.wr.wr; in rdma_rw_ctx_wrs() 913 return first_wr; in rdma_rw_ctx_wrs() 934 struct ib_send_wr *first_wr; in rdma_rw_ctx_post() local 936 first_wr = rdma_rw_ctx_wrs(ctx, qp, port_num, cqe, chain_wr); in rdma_rw_ctx_post() 937 return ib_post_send(qp, first_wr, NULL); in rdma_rw_ctx_post()
|
| /linux/drivers/infiniband/ulp/iser/ |
| H A D | iser_verbs.c | 860 struct ib_send_wr *first_wr; in iser_post_send() local 875 first_wr = &tx_desc->inv_wr; in iser_post_send() 877 first_wr = &tx_desc->reg_wr.wr; in iser_post_send() 879 first_wr = wr; in iser_post_send() 881 ret = ib_post_send(ib_conn->qp, first_wr, NULL); in iser_post_send()
|
| /linux/drivers/nvme/target/ |
| H A D | rdma.c | 713 struct ib_send_wr *first_wr; in nvmet_rdma_queue_response() local 724 first_wr = rdma_rw_ctx_wrs(&rsp->rw, cm_id->qp, in nvmet_rdma_queue_response() 727 first_wr = rdma_rw_ctx_wrs(&rsp->rw, cm_id->qp, in nvmet_rdma_queue_response() 730 first_wr = &rsp->send_wr; in nvmet_rdma_queue_response() 739 if (unlikely(ib_post_send(cm_id->qp, first_wr, NULL))) { in nvmet_rdma_queue_response()
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
| H A D | chcr_ktls.c | 842 bool first_wr = ((tx_info->prev_ack == 0) && (tx_info->prev_win == 0)); in chcr_ktls_xmit_tcb_cpls() local 872 if (first_wr || tcp_seq != tx_info->prev_seq) { in chcr_ktls_xmit_tcb_cpls() 894 if (first_wr || tx_info->prev_ack != tcp_ack) { in chcr_ktls_xmit_tcb_cpls() 903 if (first_wr || tx_info->prev_win != tcp_win) { in chcr_ktls_xmit_tcb_cpls()
|