| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_qp.c | 263 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event() 568 struct mthca_qp_context *qp_context; in __mthca_modify_qp() local 578 qp_context = &qp_param->context; in __mthca_modify_qp() 581 qp_context->flags = cpu_to_be32((to_mthca_state(new_state) << 28) | in __mthca_modify_qp() 583 qp_context->flags |= cpu_to_be32(MTHCA_QP_BIT_DE); in __mthca_modify_qp() 585 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 590 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 593 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_REARM << 11); in __mthca_modify_qp() 596 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_ARMED << 11); in __mthca_modify_qp() 604 qp_context->mtu_msgmax = (IB_MTU_2048 << 5) | 11; in __mthca_modify_qp() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | security.c | 174 if (sec->qp->event_handler && sec->qp->qp_context) { in qp_to_error() 177 sec->qp->qp_context); in qp_to_error() 185 if (qp->event_handler && qp->qp_context) { in qp_to_error() 189 qp->qp_context); in qp_to_error()
|
| H A D | mad.c | 3119 static void qp_event_handler(struct ib_event *event, void *qp_context) in qp_event_handler() argument 3121 struct ib_mad_qp_info *qp_info = qp_context; in qp_event_handler() 3163 qp_init_attr.qp_context = qp_info; in create_mad_qp()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | qp.c | 464 cmd->qp_context.pri_path.grh_mylmc = params->smac_index; in mlx4_update_qp() 479 cmd->qp_context.pri_path.fl |= in mlx4_update_qp() 487 cmd->qp_context.param3 |= cpu_to_be32(MLX4_STRIP_VLAN); in mlx4_update_qp() 492 cmd->qp_context.rate_limit_params = cpu_to_be16((params->rate_unit << 14) | params->rate_val); in mlx4_update_qp() 503 cmd->qp_context.qos_vport = params->qos_vport; in mlx4_update_qp()
|
| H A D | resource_tracker.c | 4318 smac_index = cmd->qp_context.pri_path.grh_mylmc; in mlx4_UPDATE_QP_wrapper() 5263 ctx->qp_context.qos_vport = work->qos_vport; in update_qos_vpp() 5354 upd_context->qp_context.param3 = qp->param3; in mlx4_vf_immed_vlan_work_handler() 5355 upd_context->qp_context.pri_path.vlan_control = qp->vlan_control; in mlx4_vf_immed_vlan_work_handler() 5356 upd_context->qp_context.pri_path.fvl_rx = qp->fvl_rx; in mlx4_vf_immed_vlan_work_handler() 5357 upd_context->qp_context.pri_path.vlan_index = qp->vlan_index; in mlx4_vf_immed_vlan_work_handler() 5358 upd_context->qp_context.pri_path.fl = qp->pri_path_fl; in mlx4_vf_immed_vlan_work_handler() 5359 upd_context->qp_context.pri_path.feup = qp->feup; in mlx4_vf_immed_vlan_work_handler() 5360 upd_context->qp_context.pri_path.sched_queue = in mlx4_vf_immed_vlan_work_handler() 5363 upd_context->qp_context.param3 = qp->param3 & ~cpu_to_be32(MLX4_STRIP_VLAN); in mlx4_vf_immed_vlan_work_handler() [all …]
|
| /linux/drivers/infiniband/ulp/isert/ |
| H A D | ib_isert.c | 119 attr.qp_context = isert_conn; in isert_create_qp() 513 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connected_handler() 627 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_disconnected_handler() 654 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connect_error() 697 isert_conn = cma_id->qp->qp_context; in isert_cma_handler() 1316 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_recv_done() 1375 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_recv_done() 1573 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_write_done() 1615 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_read_done() 1691 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_send_done() [all …]
|
| /linux/drivers/infiniband/hw/usnic/ |
| H A D | usnic_ib_qp_grp.c | 492 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 497 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 504 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify()
|
| /linux/drivers/infiniband/ulp/iser/ |
| H A D | iser_initiator.c | 534 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_login_rsp() 635 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_task_rsp() 698 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_dataout_comp()
|
| H A D | iser_verbs.c | 261 init_attr.qp_context = (void *)ib_conn; in iser_create_ib_conn_res() 939 struct iser_conn *iser_conn = to_iser_conn(wc->qp->qp_context); in iser_err_comp()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | gsi.c | 205 .qp_context = gsi->rx_qp->qp_context, in create_gsi_ud_qp()
|
| /linux/drivers/nvme/target/ |
| H A D | rdma.c | 696 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_send_done() 749 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_read_data_done() 782 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_write_data_done() 1022 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_recv_done() 1282 qp_attr.qp_context = queue; in nvmet_rdma_create_queue_ib() 1774 queue = cm_id->qp->qp_context; in nvmet_rdma_cm_handler()
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | ev.c | 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event()
|
| /linux/net/smc/ |
| H A D | smc_wr.c | 83 link = wc->qp->qp_context; in smc_wr_tx_process_cqe() 432 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_wr_rx_demultiplex() 455 link = wc[i].qp->qp_context; in smc_wr_rx_process_cqes()
|
| H A D | smc_cdc.c | 471 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_cdc_rx_handler()
|
| H A D | smc_ib.c | 667 .qp_context = lnk, in smc_ib_create_queue_pair()
|
| /linux/include/linux/mlx4/ |
| H A D | qp.h | 240 struct mlx4_qp_context qp_context; member
|
| /linux/drivers/infiniband/ulp/ipoib/ |
| H A D | ipoib_cm.c | 262 .qp_context = p, in ipoib_cm_create_rx_qp() 589 p = wc->qp->qp_context; in ipoib_cm_handle_rx_wc() 797 struct ipoib_cm_tx *tx = wc->qp->qp_context; in ipoib_cm_handle_tx_wc() 1068 .qp_context = tx, in ipoib_cm_create_tx_qp()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_qp.c | 465 init->qp_context = qp->ibqp.qp_context; in rxe_qp_to_init()
|
| /linux/net/9p/ |
| H A D | trans_rdma.c | 598 qp_attr.qp_context = client; in rdma_create_trans()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | svc_rdma_transport.c | 500 qp_attr.qp_context = &newxprt->sc_xprt; in svc_rdma_accept()
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | rtrs.c | 266 init_attr.qp_context = con; in create_qp()
|
| /linux/drivers/nvme/host/ |
| H A D | rdma.c | 277 init_attr.qp_context = queue; in nvme_rdma_create_qp() 1170 struct nvme_rdma_queue *queue = wc->qp->qp_context; in nvme_rdma_wr_error() 1738 struct nvme_rdma_queue *queue = wc->qp->qp_context; in nvme_rdma_recv_done()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | mad.c | 1784 static void pv_qp_event_handler(struct ib_event *event, void *qp_context) in pv_qp_event_handler() argument 1786 struct mlx4_ib_demux_pv_ctx *sqp = qp_context; in pv_qp_event_handler() 1830 qp_init_attr.init_attr.qp_context = ctx; in create_pv_sqp()
|
| /linux/drivers/infiniband/hw/qedr/ |
| H A D | qedr_iw_cm.c | 197 ibqp->event_handler(&event, ibqp->qp_context); in qedr_iw_qp_event()
|
| H A D | main.c | 744 ibqp->event_handler(&event, ibqp->qp_context); in qedr_affiliated_event()
|