| /linux/drivers/infiniband/hw/mana/ |
| H A D | qp.c | 269 struct mana_ib_cq *send_cq = in mana_ib_create_qp_raw() local 270 container_of(attr->send_cq, struct mana_ib_cq, ibcq); in mana_ib_create_qp_raw() 339 cq_spec.gdma_region = send_cq->queue.gdma_region; in mana_ib_create_qp_raw() 340 cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_raw() 342 eq_vec = send_cq->comp_vector; in mana_ib_create_qp_raw() 357 send_cq->queue.gdma_region = GDMA_INVALID_DMA_REGION; in mana_ib_create_qp_raw() 360 send_cq->queue.id = cq_spec.queue_index; in mana_ib_create_qp_raw() 363 err = mana_ib_install_cq_cb(mdev, send_cq); in mana_ib_create_qp_raw() 369 qp->qp_handle, qp->raw_sq.id, send_cq->queue.id); in mana_ib_create_qp_raw() 372 resp.cqid = send_cq->queue.id; in mana_ib_create_qp_raw() [all …]
|
| H A D | main.c | 1022 struct mana_ib_cq *send_cq = container_of(qp->ibqp.send_cq, struct mana_ib_cq, ibcq); in mana_ib_gd_create_rc_qp() local 1034 req.send_cq_handle = send_cq->cq_handle; in mana_ib_gd_create_rc_qp() 1081 struct mana_ib_cq *send_cq = container_of(qp->ibqp.send_cq, struct mana_ib_cq, ibcq); in mana_ib_gd_create_ud_qp() local 1093 req.send_cq_handle = send_cq->cq_handle; in mana_ib_gd_create_ud_qp()
|
| /linux/drivers/infiniband/ulp/ipoib/ |
| H A D | ipoib_verbs.c | 187 priv->send_cq = ib_create_cq(priv->ca, ipoib_ib_tx_completion, NULL, in ipoib_transport_dev_init() 189 if (IS_ERR(priv->send_cq)) { in ipoib_transport_dev_init() 197 init_attr.send_cq = priv->send_cq; in ipoib_transport_dev_init() 218 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init() 244 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_init() 266 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_cleanup()
|
| H A D | ipoib_ib.c | 440 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in poll_tx() 508 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in ipoib_tx_poll() 520 if (unlikely(ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_tx_poll() 674 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_send()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | qp.c | 88 struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq); 827 static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq, 829 static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq, 1296 static int get_sq_ts_format(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *send_cq) in get_sq_ts_format() argument 1300 return get_ts_format(dev, send_cq, fr_supported(ts_cap), in get_sq_ts_format() 1304 static int get_qp_ts_format(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *send_cq, in get_qp_ts_format() argument 1313 send_cq ? get_ts_format(dev, send_cq, fr_sup, rt_sup) : in get_qp_ts_format() 1590 to_mcq(init_attr->send_cq)); in create_raw_packet_qp() 1931 scqe_sz = mlx5_ib_get_cqe_size(init_attr->send_cq); in configure_requester_scat_cqe() 2081 struct mlx5_ib_cq *send_cq; in create_dci() local [all …]
|
| H A D | gsi.c | 50 struct ib_cq *gsi_cq = mqp->ibqp.send_cq; in generate_completions() 141 hw_init_attr.send_cq = gsi->cq; in mlx5_ib_create_gsi() 206 .send_cq = gsi->cq, in create_gsi_ud_qp()
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_qp.c | 735 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); in __mthca_modify_qp() 837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp() 838 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL); in __mthca_modify_qp() 1164 struct mthca_cq *send_cq, in mthca_alloc_qp_common() argument 1293 struct mthca_cq *send_cq, in mthca_alloc_qp() argument 1321 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, in mthca_alloc_qp() 1336 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) in mthca_lock_cqs() argument 1337 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mthca_lock_cqs() 1339 if (send_cq == recv_cq) { in mthca_lock_cqs() 1340 spin_lock_irq(&send_cq->lock); in mthca_lock_cqs() [all …]
|
| H A D | mthca_dev.h | 535 struct mthca_cq *send_cq, 544 struct mthca_cq *send_cq,
|
| /linux/include/rdma/ |
| H A D | rdmavt_qp.h | 808 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); in rvt_send_cq() 923 static inline u32 ib_cq_tail(struct ib_cq *send_cq) in ib_cq_tail() argument 925 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_tail() 927 return ibcq_to_rvtcq(send_cq)->ip ? in ib_cq_tail() 929 ibcq_to_rvtcq(send_cq)->kqueue->tail; in ib_cq_tail() 939 static inline u32 ib_cq_head(struct ib_cq *send_cq) in ib_cq_head() argument 941 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_head() 943 return ibcq_to_rvtcq(send_cq)->ip ? in ib_cq_head() 945 ibcq_to_rvtcq(send_cq)->kqueue->head; in ib_cq_head()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | qp.c | 52 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, 54 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, 760 if (init_attr->send_cq || init_attr->cap.max_send_wr) { in _mlx4_ib_create_qp_rss() 962 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_rq() 971 mcq = to_mcq(init_attr->send_cq); in create_rq() 975 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), in create_rq() 1240 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_qp_common() 1249 mcq = to_mcq(init_attr->send_cq); in create_qp_common() 1253 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), in create_qp_common() 1308 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) in mlx4_ib_lock_cqs() argument [all …]
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | verbs.c | 350 if (ep->re_attr.send_cq) in rpcrdma_ep_destroy() 351 ib_free_cq(ep->re_attr.send_cq); in rpcrdma_ep_destroy() 352 ep->re_attr.send_cq = NULL; in rpcrdma_ep_destroy() 421 ep->re_attr.send_cq = ib_alloc_cq_any(device, r_xprt, in rpcrdma_ep_create() 424 if (IS_ERR(ep->re_attr.send_cq)) { in rpcrdma_ep_create() 425 rc = PTR_ERR(ep->re_attr.send_cq); in rpcrdma_ep_create() 426 ep->re_attr.send_cq = NULL; in rpcrdma_ep_create() 624 sc->sc_cid.ci_queue_id = ep->re_attr.send_cq->res.id; in rpcrdma_sendctx_create()
|
| /linux/Documentation/translations/zh_CN/infiniband/ |
| H A D | tag_matching.rst | 39 1. Eager协议--当发送方处理完发送时,完整的信息就会被发送。在send_cq中会收到
|
| /linux/drivers/infiniband/hw/ionic/ |
| H A D | ionic_controlpath.c | 1320 struct ionic_cq *send_cq, in ionic_create_qp_cmd() argument 1353 wr.wqe.cmd.create_qp.sq_cq_id = cpu_to_le32(send_cq->cqid); in ionic_create_qp_cmd() 2184 udma_mask &= to_ionic_vcq(attr->send_cq)->udma_mask; in ionic_create_qp() 2236 to_ionic_vcq_cq(attr->send_cq, qp->udma_idx), in ionic_create_qp() 2340 cq = to_ionic_vcq_cq(attr->send_cq, qp->udma_idx); in ionic_create_qp() 2400 if (qp->ibqp.send_cq) in ionic_notify_qp_cqs() 2401 ionic_notify_flush_cq(to_ionic_vcq_cq(qp->ibqp.send_cq, in ionic_notify_qp_cqs() 2403 if (qp->ibqp.recv_cq && qp->ibqp.recv_cq != qp->ibqp.send_cq) in ionic_notify_qp_cqs() 2413 if (qp->ibqp.send_cq) { in ionic_flush_qp() 2414 cq = to_ionic_vcq_cq(qp->ibqp.send_cq, qp->udma_idx); in ionic_flush_qp() [all …]
|
| /linux/drivers/net/ethernet/ibm/ehea/ |
| H A D | ehea_main.c | 199 arr[i++].fwh = pr->send_cq->fw_handle; in ehea_update_firmware_handles() 804 struct ehea_cq *send_cq = pr->send_cq; in ehea_proc_cqes() local 813 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes() 815 ehea_inc_cq(send_cq); in ehea_proc_cqes() 855 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes() 858 ehea_update_feca(send_cq, cqe_counter); in ehea_proc_cqes() 893 ehea_reset_cq_ep(pr->send_cq); in ehea_poll() 895 ehea_reset_cq_n1(pr->send_cq); in ehea_poll() 898 cqe_skb = ehea_poll_cq(pr->send_cq); in ehea_poll() 1477 pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq, in ehea_init_port_res() [all …]
|
| H A D | ehea.h | 351 struct ehea_cq *send_cq; member
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_qp.c | 58 static inline void get_cqs(struct pvrdma_qp *qp, struct pvrdma_cq **send_cq, in get_cqs() argument 61 *send_cq = to_vcq(qp->ibqp.send_cq); in get_cqs() 359 cmd->send_cq_handle = to_vcq(init_attr->send_cq)->cq_handle; in pvrdma_create_qp() 1034 init_attr->send_cq = qp->ibqp.send_cq; in pvrdma_query_qp()
|
| /linux/fs/smb/server/ |
| H A D | transport_rdma.c | 516 if (sc->ib.send_cq) in free_transport() 517 ib_free_cq(sc->ib.send_cq); in free_transport() 2435 sc->ib.send_cq = ib_alloc_cq_any(sc->ib.dev, sc, in smb_direct_create_qpair() 2438 if (IS_ERR(sc->ib.send_cq)) { in smb_direct_create_qpair() 2440 ret = PTR_ERR(sc->ib.send_cq); in smb_direct_create_qpair() 2441 sc->ib.send_cq = NULL; in smb_direct_create_qpair() 2469 qp_attr.send_cq = sc->ib.send_cq; in smb_direct_create_qpair() 2492 if (sc->ib.send_cq) { in smb_direct_create_qpair() 2493 ib_destroy_cq(sc->ib.send_cq); in smb_direct_create_qpair() 2494 sc->ib.send_cq = NULL; in smb_direct_create_qpair()
|
| /linux/drivers/infiniband/core/ |
| H A D | verbs.c | 1245 qp->send_cq = qp->recv_cq = NULL; in create_xrc_qp_user() 1299 qp->send_cq = attr->send_cq; in create_qp() 1313 qp->send_cq = attr->send_cq; in create_qp() 1373 if (qp->send_cq) in ib_qp_usecnt_inc() 1374 atomic_inc(&qp->send_cq->usecnt); in ib_qp_usecnt_inc() 1392 if (qp->send_cq) in ib_qp_usecnt_dec() 1393 atomic_dec(&qp->send_cq->usecnt); in ib_qp_usecnt_dec() 2892 struct ib_cq *cq = qp->send_cq; in __ib_drain_sq() 3051 trace_cq_drain_complete(qp->send_cq); in ib_drain_sq()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_qp.c | 123 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init() 405 struct rxe_cq *scq = to_rcq(init->send_cq); in rxe_qp_from_init() 466 init->send_cq = qp->ibqp.send_cq; in rxe_qp_to_init()
|
| /linux/drivers/infiniband/ulp/srp/ |
| H A D | ib_srp.c | 517 ib_process_cq_direct(ch->send_cq, -1); in srp_destroy_qp() 530 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local 548 send_cq = ib_alloc_cq(dev->dev, ch, m * target->queue_size, in srp_create_ch_ib() 550 if (IS_ERR(send_cq)) { in srp_create_ch_ib() 551 ret = PTR_ERR(send_cq); in srp_create_ch_ib() 562 init_attr->send_cq = send_cq; in srp_create_ch_ib() 600 if (ch->send_cq) in srp_create_ch_ib() 601 ib_free_cq(ch->send_cq); in srp_create_ch_ib() 605 ch->send_cq = send_cq; in srp_create_ch_ib() 623 ib_free_cq(send_cq); in srp_create_ch_ib() [all …]
|
| H A D | ib_srp.h | 157 struct ib_cq *send_cq; member
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | send.c | 603 struct mlx5hws_send_ring_cq *cq = &send_ring->send_cq; in hws_send_engine_poll_cq() 990 hws_send_ring_close_cq(&queue->send_ring.send_cq); in hws_send_ring_close() 1000 err = hws_send_ring_open_cq(ctx->mdev, queue, numa_node, &ring->send_cq); in mlx5hws_send_ring_open() 1005 &ring->send_cq); in mlx5hws_send_ring_open() 1012 hws_send_ring_close_cq(&ring->send_cq); in mlx5hws_send_ring_open()
|
| /linux/fs/smb/client/ |
| H A D | smbdirect.c | 1953 ib_free_cq(sc->ib.send_cq); in smbd_destroy() 2192 sc->ib.send_cq = in _smbd_get_connection() 2195 if (IS_ERR(sc->ib.send_cq)) { in _smbd_get_connection() 2196 sc->ib.send_cq = NULL; in _smbd_get_connection() 2214 qp_attr.send_cq = sc->ib.send_cq; in _smbd_get_connection() 2321 if (sc->ib.send_cq) in _smbd_get_connection() 2322 ib_free_cq(sc->ib.send_cq); in _smbd_get_connection()
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_hw_v2.c | 720 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in hns_roce_v2_post_send() 1008 struct ib_cq *cq = ibqp->send_cq; in hns_roce_v2_drain_sq() 2964 free_mr->rsv_qp[i]->ibqp.send_cq = cq; in free_mr_init_qp() 3022 qp_init_attr.send_cq = cq; in free_mr_alloc_res() 4620 hr_reg_write(context, QPC_TX_CQN, get_cqn(ibqp->send_cq)); in modify_qp_reset_to_init() 4648 hr_reg_write(context, QPC_TX_CQN, get_cqn(ibqp->send_cq)); in modify_qp_init_to_init() 5570 if (ibqp->send_cq) in clear_qp() 5571 hns_roce_v2_cq_clean(to_hr_cq(ibqp->send_cq), in clear_qp() 5574 if (ibqp->recv_cq && ibqp->recv_cq != ibqp->send_cq) in clear_qp() 5882 qp_init_attr->send_cq = ibqp->send_cq; in hns_roce_v2_query_qp() [all …]
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw_verbs.c | 361 if (!attrs->send_cq || (!attrs->recv_cq && !attrs->srq)) { in siw_create_qp() 405 qp->scq = to_siw_cq(attrs->send_cq); in siw_create_qp() 542 qp_init_attr->send_cq = base_qp->send_cq; in siw_query_qp()
|