| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | cq.c | 173 struct mlx4_cq_context *cq_context; in mlx4_cq_modify() local 180 cq_context = mailbox->buf; in mlx4_cq_modify() 181 cq_context->cq_max_count = cpu_to_be16(count); in mlx4_cq_modify() 182 cq_context->cq_period = cpu_to_be16(period); in mlx4_cq_modify() 195 struct mlx4_cq_context *cq_context; in mlx4_cq_resize() local 203 cq_context = mailbox->buf; in mlx4_cq_resize() 204 cq_context->logsize_usrpage = cpu_to_be32(ilog2(entries) << 24); in mlx4_cq_resize() 205 cq_context->log_page_size = mtt->page_shift - 12; in mlx4_cq_resize() 207 cq_context->mtt_base_addr_h = mtt_addr >> 32; in mlx4_cq_resize() 208 cq_context->mtt_base_addr_l = cpu_to_be32(mtt_addr & 0xffffffff); in mlx4_cq_resize() [all …]
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_cq.c | 230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 773 struct mthca_cq_context *cq_context; in mthca_init_cq() local 811 cq_context = mailbox->buf; in mthca_init_cq() 824 memset(cq_context, 0, sizeof *cq_context); in mthca_init_cq() 825 cq_context->flags = cpu_to_be32(MTHCA_CQ_STATUS_OK | in mthca_init_cq() 828 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq() 830 cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index); in mthca_init_cq() 832 cq_context->logsize_usrpage |= cpu_to_be32(dev->driver_uar.index); in mthca_init_cq() 833 cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); in mthca_init_cq() [all …]
|
| /linux/net/smc/ |
| H A D | smc_wr.h | 129 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context); 134 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
|
| H A D | smc_wr.c | 162 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_tx_cq_handler() argument 164 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_tx_cq_handler() 504 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_rx_cq_handler() argument 506 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_rx_cq_handler()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_cq.c | 102 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post() 116 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_cq_post()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | frwr_ops.c | 367 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg() 463 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv() 483 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_wake() 586 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_done()
|
| H A D | svc_rdma_rw.c | 281 struct svcxprt_rdma *rdma = cq->cq_context; in svc_rdma_reply_done() 306 struct svcxprt_rdma *rdma = cq->cq_context; in svc_rdma_write_done() 340 struct svcxprt_rdma *rdma = cq->cq_context; in svc_rdma_wc_read_done()
|
| H A D | verbs.c | 147 struct rpcrdma_xprt *r_xprt = cq->cq_context; in rpcrdma_wc_send() 166 struct rpcrdma_xprt *r_xprt = cq->cq_context; in rpcrdma_wc_receive()
|
| /linux/drivers/infiniband/hw/ionic/ |
| H A D | ionic_admin.c | 492 static void ionic_rdma_admincq_comp(struct ib_cq *ibcq, void *cq_context) in ionic_rdma_admincq_comp() argument 494 struct ionic_aq *aq = cq_context; in ionic_rdma_admincq_comp() 504 static void ionic_rdma_admincq_event(struct ib_event *event, void *cq_context) in ionic_rdma_admincq_event() argument 506 struct ionic_aq *aq = cq_context; in ionic_rdma_admincq_event() 811 cq->vcq->ibcq.cq_context); in ionic_cq_event() 821 cq->vcq->ibcq.cq_context); in ionic_cq_event() 1173 vcq->ibcq.cq_context = aq; in ionic_create_rdma_admin()
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | ev.c | 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 234 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
| H A D | qp.c | 1009 schp->ibcq.cq_context); in complete_sq_drain_wr() 1060 rchp->ibcq.cq_context); in complete_rq_drain_wr() 1646 rchp->ibcq.cq_context); in __flush_qp() 1653 rchp->ibcq.cq_context); in __flush_qp() 1659 schp->ibcq.cq_context); in __flush_qp() 1683 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp() 1689 schp->ibcq.cq_context); in flush_qp()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | cq.c | 111 int eqn = MLX5_GET(cqc, MLX5_ADDR_OF(create_cq_in, in, cq_context), in mlx5_create_cq() 246 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_core_modify_cq_moderation()
|
| H A D | wc.c | 94 cqc = MLX5_ADDR_OF(create_cq_in, in, cq_context); in create_wc_cq()
|
| /linux/drivers/infiniband/core/ |
| H A D | uverbs_std_types_cq.c | 43 struct ib_uverbs_event_queue *ev_queue = cq->cq_context; in uverbs_free_cq() 188 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL; in UVERBS_HANDLER()
|
| H A D | cq.c | 228 cq->cq_context = private; in __ib_alloc_cq()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | cq.c | 48 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 68 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 790 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_user() 921 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_kernel() 949 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in notify_soft_wc_handler() 1011 cqc = MLX5_ADDR_OF(create_cq_in, cqb, cq_context); in mlx5_ib_create_cq() 1372 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_ib_resize_cq()
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_hw_v2.c | 957 ibcq->comp_handler(ibcq, ibcq->cq_context); in handle_drain_completion() 2932 free_mr->rsv_cq->ib_cq.cq_context = NULL; in free_mr_init_cq() 3861 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_write_cqc() local 3863 cq_context = mb_buf; in hns_roce_v2_write_cqc() 3864 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v2_write_cqc() 3866 hr_reg_write(cq_context, CQC_CQ_ST, V2_CQ_STATE_VALID); in hns_roce_v2_write_cqc() 3867 hr_reg_write(cq_context, CQC_ARM_ST, NO_ARMED); in hns_roce_v2_write_cqc() 3868 hr_reg_write(cq_context, CQC_SHIFT, ilog2(hr_cq->cq_depth)); in hns_roce_v2_write_cqc() 3869 hr_reg_write(cq_context, CQC_CEQN, hr_cq->vector); in hns_roce_v2_write_cqc() 3870 hr_reg_write(cq_context, CQC_CQN, hr_cq->cqn); in hns_roce_v2_write_cqc() [all …]
|
| /linux/drivers/infiniband/hw/mana/ |
| H A D | cq.c | 135 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mana_ib_cq_handler() 317 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mana_drain_gsi_sqs()
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| H A D | cq.c | 75 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter() 141 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete()
|
| /linux/net/9p/ |
| H A D | trans_rdma.c | 180 struct p9_client *client = cq->cq_context; in recv_done() 231 struct p9_client *client = cq->cq_context; in send_done()
|
| /linux/drivers/infiniband/hw/qedr/ |
| H A D | qedr_roce_cm.c | 93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet() 122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet()
|
| H A D | main.c | 474 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler() 728 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | cq.c | 46 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp() 65 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| H A D | aso.c | 94 cqc = MLX5_ADDR_OF(create_cq_in, in, cq_context); in create_aso_cq()
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_main.c | 366 ibcq->event_handler(&e, ibcq->cq_context); in pvrdma_cq_event() 519 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in pvrdma_intrx_handler()
|