Home
last modified time | relevance | path

Searched refs:recv_cq (Results 1 – 25 of 36) sorted by relevance

12

/linux/drivers/infiniband/ulp/ipoib/
H A Dipoib_verbs.c178 priv->recv_cq = ib_create_cq(priv->ca, ipoib_ib_rx_completion, NULL, in ipoib_transport_dev_init()
180 if (IS_ERR(priv->recv_cq)) { in ipoib_transport_dev_init()
194 if (ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init()
198 init_attr.recv_cq = priv->recv_cq; in ipoib_transport_dev_init()
247 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_init()
267 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_cleanup()
H A Dipoib_ib.c467 n = ib_poll_cq(priv->recv_cq, t, priv->ibwc); in ipoib_rx_poll()
489 if (unlikely(ib_req_notify_cq(priv->recv_cq, in ipoib_rx_poll()
872 ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP); in ipoib_ib_dev_stop_default()
1005 n = ib_poll_cq(priv->recv_cq, IPOIB_NUM_WC, priv->ibwc); in ipoib_drain_cq()
H A Dipoib_cm.c255 .send_cq = priv->recv_cq, /* For drain WR */ in ipoib_cm_create_rx_qp()
256 .recv_cq = priv->recv_cq, in ipoib_cm_create_rx_qp()
1062 .recv_cq = priv->recv_cq, in ipoib_cm_create_tx_qp()
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c774 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); in __mthca_modify_qp()
835 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn, in __mthca_modify_qp()
837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp()
1165 struct mthca_cq *recv_cq, in mthca_alloc_qp_common() argument
1294 struct mthca_cq *recv_cq, in mthca_alloc_qp() argument
1321 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, in mthca_alloc_qp()
1336 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) in mthca_lock_cqs() argument
1337 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mthca_lock_cqs()
1339 if (send_cq == recv_cq) { in mthca_lock_cqs()
1341 __acquire(&recv_cq->lock); in mthca_lock_cqs()
[all …]
H A Dmthca_dev.h536 struct mthca_cq *recv_cq,
545 struct mthca_cq *recv_cq,
H A Dmthca_provider.c505 to_mcq(init_attr->recv_cq), in mthca_create_qp()
530 to_mcq(init_attr->recv_cq), in mthca_create_qp()
/linux/net/sunrpc/xprtrdma/
H A Dverbs.c347 if (ep->re_attr.recv_cq) in rpcrdma_ep_destroy()
348 ib_free_cq(ep->re_attr.recv_cq); in rpcrdma_ep_destroy()
349 ep->re_attr.recv_cq = NULL; in rpcrdma_ep_destroy()
430 ep->re_attr.recv_cq = ib_alloc_cq_any(device, r_xprt, in rpcrdma_ep_create()
433 if (IS_ERR(ep->re_attr.recv_cq)) { in rpcrdma_ep_create()
434 rc = PTR_ERR(ep->re_attr.recv_cq); in rpcrdma_ep_create()
435 ep->re_attr.recv_cq = NULL; in rpcrdma_ep_create()
1395 rep->rr_cid.ci_queue_id = ep->re_attr.recv_cq->res.id; in rpcrdma_post_recvs()
H A Dsvc_rdma_transport.c510 qp_attr.recv_cq = newxprt->sc_rq_cq; in svc_rdma_accept()
/linux/drivers/infiniband/hw/ionic/
H A Dionic_controlpath.c1323 struct ionic_cq *recv_cq, in ionic_create_qp_cmd() argument
1366 wr.wqe.cmd.create_qp.rq_cq_id = cpu_to_le32(recv_cq->cqid); in ionic_create_qp_cmd()
2189 udma_mask &= to_ionic_vcq(attr->recv_cq)->udma_mask; in ionic_create_qp()
2239 to_ionic_vcq_cq(attr->recv_cq, qp->udma_idx), in ionic_create_qp()
2357 cq = to_ionic_vcq_cq(attr->recv_cq, qp->udma_idx); in ionic_create_qp()
2405 if (qp->ibqp.recv_cq && qp->ibqp.recv_cq != qp->ibqp.send_cq) in ionic_notify_qp_cqs()
2406 ionic_notify_flush_cq(to_ionic_vcq_cq(qp->ibqp.recv_cq, in ionic_notify_qp_cqs()
2430 if (qp->ibqp.recv_cq) { in ionic_flush_qp()
2431 cq = to_ionic_vcq_cq(qp->ibqp.recv_cq, qp->udma_idx); in ionic_flush_qp()
2488 if (qp->ibqp.recv_cq) { in ionic_reset_qp()
[all …]
/linux/fs/smb/smbdirect/
H A Dconnection.c357 sc->ib.recv_cq = ib_alloc_cq_any(sc->ib.dev, sc, in smbdirect_connection_create_qp()
360 if (IS_ERR(sc->ib.recv_cq)) { in smbdirect_connection_create_qp()
361 pr_err("Can't create RDMA recv CQ: %1pe\n", sc->ib.recv_cq); in smbdirect_connection_create_qp()
362 ret = PTR_ERR(sc->ib.recv_cq); in smbdirect_connection_create_qp()
363 sc->ib.recv_cq = NULL; in smbdirect_connection_create_qp()
382 qp_attr.recv_cq = sc->ib.recv_cq; in smbdirect_connection_create_qp()
406 if (sc->ib.recv_cq) { in smbdirect_connection_destroy_qp()
407 ib_destroy_cq(sc->ib.recv_cq); in smbdirect_connection_destroy_qp()
408 sc->ib.recv_cq = NULL; in smbdirect_connection_destroy_qp()
/linux/drivers/infiniband/sw/rxe/
H A Drxe_qp.c123 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init()
404 struct rxe_cq *rcq = to_rcq(init->recv_cq); in rxe_qp_from_init()
467 init->recv_cq = qp->ibqp.recv_cq; in rxe_qp_to_init()
/linux/drivers/infiniband/ulp/srp/
H A Dib_srp.h158 struct ib_cq *recv_cq; member
H A Dib_srp.c531 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local
542 recv_cq = ib_alloc_cq(dev->dev, ch, target->queue_size + 1, in srp_create_ch_ib()
544 if (IS_ERR(recv_cq)) { in srp_create_ch_ib()
545 ret = PTR_ERR(recv_cq); in srp_create_ch_ib()
564 init_attr->recv_cq = recv_cq; in srp_create_ch_ib()
599 if (ch->recv_cq) in srp_create_ch_ib()
600 ib_free_cq(ch->recv_cq); in srp_create_ch_ib()
605 ch->recv_cq = recv_cq; in srp_create_ch_ib()
627 ib_free_cq(recv_cq); in srp_create_ch_ib()
670 ib_free_cq(ch->recv_cq); in srp_free_ch_ib()
[all …]
/linux/drivers/infiniband/hw/mlx5/
H A Dgsi.c207 .recv_cq = gsi->rx_qp->recv_cq, in create_gsi_ud_qp()
/linux/drivers/net/ethernet/ibm/ehea/
H A Dehea_main.c201 arr[i++].fwh = pr->recv_cq->fw_handle; in ehea_update_firmware_handles()
892 ehea_reset_cq_ep(pr->recv_cq); in ehea_poll()
894 ehea_reset_cq_n1(pr->recv_cq); in ehea_poll()
1469 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, in ehea_init_port_res()
1472 if (!pr->recv_cq) { in ehea_init_port_res()
1488 pr->recv_cq->attr.act_nr_of_cqes); in ehea_init_port_res()
1513 init_attr->recv_cq_handle = pr->recv_cq->fw_handle; in ehea_init_port_res()
1563 ehea_destroy_cq(pr->recv_cq); in ehea_init_port_res()
1580 ehea_destroy_cq(pr->recv_cq); in ehea_clean_portres()
H A Dehea.h352 struct ehea_cq *recv_cq; member
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_hw_v2.c849 hr_qp->ibqp.recv_cq))) { in hns_roce_v2_post_recv()
971 struct ib_cq *cq = ibqp->recv_cq; in hns_roce_v2_drain_rq()
2963 free_mr->rsv_qp[i]->ibqp.recv_cq = cq; in free_mr_init_qp()
3023 qp_init_attr.recv_cq = cq; in free_mr_alloc_res()
4611 hr_reg_write(context, QPC_RX_CQN, get_cqn(ibqp->recv_cq)); in modify_qp_reset_to_init()
4645 hr_reg_write(context, QPC_RX_CQN, get_cqn(ibqp->recv_cq)); in modify_qp_init_to_init()
5574 if (ibqp->recv_cq && ibqp->recv_cq != ibqp->send_cq) in clear_qp()
5575 hns_roce_v2_cq_clean(to_hr_cq(ibqp->recv_cq), in clear_qp()
5881 qp_init_attr->recv_cq = ibqp->recv_cq; in hns_roce_v2_query_qp()
5906 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v2_destroy_qp_common() local
[all …]
/linux/drivers/infiniband/hw/qedr/
H A Dqedr_roce_cm.c63 dev->gsi_rqcq = get_qedr_cq(attrs->recv_cq); in qedr_store_gsi_qp_cq()
144 cq = get_qedr_cq(attrs->recv_cq); in qedr_destroy_gsi_cq()
/linux/net/9p/
H A Dtrans_rdma.c606 qp_attr.recv_cq = rdma->cq; in rdma_create_trans()
/linux/include/rdma/
H A Drdmavt_qp.h788 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.recv_cq); in rvt_recv_cq()
/linux/drivers/infiniband/ulp/rtrs/
H A Drtrs.c271 init_attr.recv_cq = con->cq; in create_qp()
/linux/include/trace/events/
H A Drpcrdma.h990 __entry->cq_id = ep ? ep->re_attr.recv_cq->res.id : 0;
1024 __entry->cq_id = ep->re_attr.recv_cq->res.id;
1055 __entry->cq_id = ep->re_attr.recv_cq->res.id;
/linux/drivers/infiniband/hw/efa/
H A Defa_verbs.c391 qp_init_attr->recv_cq = ibqp->recv_cq; in efa_query_qp()
741 create_qp_params.recv_cq_idx = to_ecq(init_attr->recv_cq)->cq_idx; in efa_create_qp()
/linux/drivers/infiniband/ulp/iser/
H A Diser_verbs.c263 init_attr.recv_cq = ib_conn->cq; in iser_create_ib_conn_res()
/linux/drivers/infiniband/hw/cxgb4/
H A Dqp.c1040 rchp = to_c4iw_cq(qhp->ibqp.recv_cq); in complete_rq_drain_wr()
1670 rchp = to_c4iw_cq(qhp->ibqp.recv_cq); in flush_qp()
2137 rchp = get_chp(rhp, ((struct c4iw_cq *)attrs->recv_cq)->cq.cqid); in c4iw_create_qp()
2194 qhp->attr.rcq = ((struct c4iw_cq *) attrs->recv_cq)->cq.cqid; in c4iw_create_qp()

12