Home
last modified time | relevance | path

Searched refs:cq_context (Results 1 – 25 of 34) sorted by relevance

12

/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dcq.c173 struct mlx4_cq_context *cq_context; in mlx4_cq_modify() local
180 cq_context = mailbox->buf; in mlx4_cq_modify()
181 cq_context->cq_max_count = cpu_to_be16(count); in mlx4_cq_modify()
182 cq_context->cq_period = cpu_to_be16(period); in mlx4_cq_modify()
195 struct mlx4_cq_context *cq_context; in mlx4_cq_resize() local
203 cq_context = mailbox->buf; in mlx4_cq_resize()
204 cq_context->logsize_usrpage = cpu_to_be32(ilog2(entries) << 24); in mlx4_cq_resize()
205 cq_context->log_page_size = mtt->page_shift - 12; in mlx4_cq_resize()
207 cq_context->mtt_base_addr_h = mtt_addr >> 32; in mlx4_cq_resize()
208 cq_context->mtt_base_addr_l = cpu_to_be32(mtt_addr & 0xffffffff); in mlx4_cq_resize()
[all …]
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_cq.c230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion()
256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event()
773 struct mthca_cq_context *cq_context; in mthca_init_cq() local
811 cq_context = mailbox->buf; in mthca_init_cq()
824 memset(cq_context, 0, sizeof *cq_context); in mthca_init_cq()
825 cq_context->flags = cpu_to_be32(MTHCA_CQ_STATUS_OK | in mthca_init_cq()
828 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq()
830 cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index); in mthca_init_cq()
832 cq_context->logsize_usrpage |= cpu_to_be32(dev->driver_uar.index); in mthca_init_cq()
833 cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); in mthca_init_cq()
[all …]
/linux/net/smc/
H A Dsmc_wr.h129 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
134 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
/linux/drivers/infiniband/sw/rxe/
H A Drxe_cq.c102 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
116 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_cq_post()
/linux/net/sunrpc/xprtrdma/
H A Dfrwr_ops.c367 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg()
463 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv()
483 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_wake()
586 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_done()
H A Dsvc_rdma_rw.c260 struct svcxprt_rdma *rdma = cq->cq_context; in svc_rdma_reply_done()
285 struct svcxprt_rdma *rdma = cq->cq_context; in svc_rdma_write_done()
319 struct svcxprt_rdma *rdma = cq->cq_context; in svc_rdma_wc_read_done()
H A Dverbs.c147 struct rpcrdma_xprt *r_xprt = cq->cq_context; in rpcrdma_wc_send()
166 struct rpcrdma_xprt *r_xprt = cq->cq_context; in rpcrdma_wc_receive()
H A Dsvc_rdma_sendto.c307 struct svcxprt_rdma *rdma = cq->cq_context; in svc_rdma_wc_send()
/linux/drivers/infiniband/hw/cxgb4/
H A Dev.c115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event()
234 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Dcq.c111 int eqn = MLX5_GET(cqc, MLX5_ADDR_OF(create_cq_in, in, cq_context), in mlx5_create_cq()
246 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_core_modify_cq_moderation()
H A Dwc.c93 cqc = MLX5_ADDR_OF(create_cq_in, in, cq_context); in create_wc_cq()
H A Ddebugfs.c441 ctx = MLX5_ADDR_OF(query_cq_out, out, cq_context); in cq_read_field()
/linux/drivers/infiniband/core/
H A Duverbs_std_types_cq.c43 struct ib_uverbs_event_queue *ev_queue = cq->cq_context; in uverbs_free_cq()
188 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL; in UVERBS_HANDLER()
H A Dverbs.c2139 void *cq_context, in __ib_create_cq() argument
2154 cq->cq_context = cq_context; in __ib_create_cq()
/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c48 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp()
68 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event()
790 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_user()
921 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_kernel()
949 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in notify_soft_wc_handler()
1011 cqc = MLX5_ADDR_OF(create_cq_in, cqb, cq_context); in mlx5_ib_create_cq()
1372 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_ib_resize_cq()
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_hw_v2.c2768 free_mr->rsv_cq->ib_cq.cq_context = NULL; in free_mr_init_cq()
3697 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_write_cqc() local
3699 cq_context = mb_buf; in hns_roce_v2_write_cqc()
3700 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v2_write_cqc()
3702 hr_reg_write(cq_context, CQC_CQ_ST, V2_CQ_STATE_VALID); in hns_roce_v2_write_cqc()
3703 hr_reg_write(cq_context, CQC_ARM_ST, NO_ARMED); in hns_roce_v2_write_cqc()
3704 hr_reg_write(cq_context, CQC_SHIFT, ilog2(hr_cq->cq_depth)); in hns_roce_v2_write_cqc()
3705 hr_reg_write(cq_context, CQC_CEQN, hr_cq->vector); in hns_roce_v2_write_cqc()
3706 hr_reg_write(cq_context, CQC_CQN, hr_cq->cqn); in hns_roce_v2_write_cqc()
3709 hr_reg_write(cq_context, CQC_CQE_SIZE, CQE_SIZE_64B); in hns_roce_v2_write_cqc()
[all …]
H A Dhns_roce_cq.c522 ibcq->comp_handler(ibcq, ibcq->cq_context); in hns_roce_cq_completion()
556 ibcq->event_handler(&event, ibcq->cq_context); in hns_roce_cq_event()
/linux/drivers/infiniband/sw/rdmavt/
H A Dcq.c75 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter()
141 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete()
/linux/net/9p/
H A Dtrans_rdma.c180 struct p9_client *client = cq->cq_context; in recv_done()
231 struct p9_client *client = cq->cq_context; in send_done()
/linux/drivers/infiniband/hw/qedr/
H A Dqedr_roce_cm.c93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet()
122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet()
H A Dmain.c477 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler()
731 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
/linux/include/rdma/
H A Dib_verbs.h1601 typedef void (*ib_comp_handler)(struct ib_cq *cq, void *cq_context);
1617 void *cq_context; member
4028 void *cq_context,
/linux/drivers/infiniband/hw/ocrdma/
H A Docrdma_hw.c788 cq->ibcq.event_handler(&ib_evt, cq->ibcq.cq_context); in ocrdma_dispatch_ibevent()
960 (*bcq->ibcq.comp_handler) (&bcq->ibcq, bcq->ibcq.cq_context); in ocrdma_qp_buddy_cq_handler()
979 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in ocrdma_qp_cq_handler()
/linux/drivers/infiniband/hw/bnxt_re/
H A Dmain.c1660 cq->ib_cq.event_handler(&ibevent, cq->ib_cq.cq_context); in bnxt_re_handle_cq_async_error()
1743 (*cq->ib_cq.comp_handler)(&cq->ib_cq, cq->ib_cq.cq_context); in bnxt_re_cqn_handler()
/linux/drivers/infiniband/ulp/srp/
H A Dib_srp.c1871 struct srp_rdma_ch *ch = cq->cq_context; in srp_send_done()
2062 struct srp_rdma_ch *ch = cq->cq_context; in srp_recv_done()
2138 struct srp_rdma_ch *ch = cq->cq_context; in srp_handle_qp_err()

12