/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 263 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event() 568 struct mthca_qp_context *qp_context; in __mthca_modify_qp() local 578 qp_context = &qp_param->context; in __mthca_modify_qp() 581 qp_context->flags = cpu_to_be32((to_mthca_state(new_state) << 28) | in __mthca_modify_qp() 583 qp_context->flags |= cpu_to_be32(MTHCA_QP_BIT_DE); in __mthca_modify_qp() 585 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 590 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_MIGRATED << 11); in __mthca_modify_qp() 593 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_REARM << 11); in __mthca_modify_qp() 596 qp_context->flags |= cpu_to_be32(MTHCA_QP_PM_ARMED << 11); in __mthca_modify_qp() 604 qp_context->mtu_msgmax = (IB_MTU_2048 << 5) | 11; in __mthca_modify_qp() [all …]
|
/linux/drivers/infiniband/core/ |
H A D | security.c | 174 if (sec->qp->event_handler && sec->qp->qp_context) { in qp_to_error() 177 sec->qp->qp_context); in qp_to_error() 185 if (qp->event_handler && qp->qp_context) { in qp_to_error() 189 qp->qp_context); in qp_to_error()
|
H A D | verbs.c | 1111 qp->registered_event_handler(event, qp->qp_context); in __ib_qp_event_handler() 1122 event->element.qp->event_handler(event, event->element.qp->qp_context); in __ib_shared_qp_event_handler() 1128 void *qp_context) in __ib_open_qp() argument 1149 qp->qp_context = qp_context; in __ib_open_qp() 1175 qp_open_attr->qp_context); in ib_open_qp() 1188 qp->qp_context = qp; in create_xrc_qp_user() 1197 qp_init_attr->qp_context); in create_xrc_qp_user() 1237 qp->qp_context = attr->qp_context; in create_qp()
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | qp.c | 464 cmd->qp_context.pri_path.grh_mylmc = params->smac_index; in mlx4_update_qp() 479 cmd->qp_context.pri_path.fl |= in mlx4_update_qp() 487 cmd->qp_context.param3 |= cpu_to_be32(MLX4_STRIP_VLAN); in mlx4_update_qp() 492 cmd->qp_context.rate_limit_params = cpu_to_be16((params->rate_unit << 14) | params->rate_val); in mlx4_update_qp() 503 cmd->qp_context.qos_vport = params->qos_vport; in mlx4_update_qp()
|
H A D | resource_tracker.c | 4325 smac_index = cmd->qp_context.pri_path.grh_mylmc; in mlx4_UPDATE_QP_wrapper() 5271 ctx->qp_context.qos_vport = work->qos_vport; in update_qos_vpp() 5362 upd_context->qp_context.param3 = qp->param3; in mlx4_vf_immed_vlan_work_handler() 5363 upd_context->qp_context.pri_path.vlan_control = qp->vlan_control; in mlx4_vf_immed_vlan_work_handler() 5364 upd_context->qp_context.pri_path.fvl_rx = qp->fvl_rx; in mlx4_vf_immed_vlan_work_handler() 5365 upd_context->qp_context.pri_path.vlan_index = qp->vlan_index; in mlx4_vf_immed_vlan_work_handler() 5366 upd_context->qp_context.pri_path.fl = qp->pri_path_fl; in mlx4_vf_immed_vlan_work_handler() 5367 upd_context->qp_context.pri_path.feup = qp->feup; in mlx4_vf_immed_vlan_work_handler() 5368 upd_context->qp_context.pri_path.sched_queue = in mlx4_vf_immed_vlan_work_handler() 5371 upd_context->qp_context.param3 = qp->param3 & ~cpu_to_be32(MLX4_STRIP_VLAN); in mlx4_vf_immed_vlan_work_handler() [all …]
|
/linux/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_qp_grp.c | 492 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 497 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 504 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify()
|
/linux/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.c | 119 attr.qp_context = isert_conn; in isert_create_qp() 515 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connected_handler() 629 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_disconnected_handler() 656 struct isert_conn *isert_conn = cma_id->qp->qp_context; in isert_connect_error() 699 isert_conn = cma_id->qp->qp_context; in isert_cma_handler() 1318 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_recv_done() 1377 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_recv_done() 1575 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_write_done() 1617 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_rdma_read_done() 1693 struct isert_conn *isert_conn = wc->qp->qp_context; in isert_login_send_done() [all …]
|
/linux/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 535 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_login_rsp() 636 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_task_rsp() 699 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_dataout_comp()
|
H A D | iser_verbs.c | 261 init_attr.qp_context = (void *)ib_conn; in iser_create_ib_conn_res() 939 struct iser_conn *iser_conn = to_iser_conn(wc->qp->qp_context); in iser_err_comp()
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | gsi.c | 204 .qp_context = gsi->rx_qp->qp_context, in create_gsi_ud_qp()
|
/linux/drivers/nvme/target/ |
H A D | rdma.c | 697 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_send_done() 750 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_read_data_done() 783 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_write_data_done() 1003 struct nvmet_rdma_queue *queue = wc->qp->qp_context; in nvmet_rdma_recv_done() 1271 qp_attr.qp_context = queue; in nvmet_rdma_create_queue_ib() 1760 queue = cm_id->qp->qp_context; in nvmet_rdma_cm_handler()
|
/linux/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event()
|
/linux/drivers/infiniband/sw/rdmavt/ |
H A D | qp.c | 717 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_qp_mr_clean() 1650 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_modify_qp() 1656 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_modify_qp() 1757 init_attr->qp_context = qp->ibqp.qp_context; in rvt_query_qp() 2459 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_comm_est() 2479 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_rc_error() 3209 sqp->ibqp.event_handler(&ev, sqp->ibqp.qp_context); in rvt_ruc_loopback()
|
/linux/net/smc/ |
H A D | smc_wr.c | 83 link = wc->qp->qp_context; in smc_wr_tx_process_cqe() 432 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_wr_rx_demultiplex() 455 link = wc[i].qp->qp_context; in smc_wr_rx_process_cqes()
|
H A D | smc_cdc.c | 471 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_cdc_rx_handler()
|
/linux/include/linux/mlx4/ |
H A D | qp.h | 240 struct mlx4_qp_context qp_context; member
|
/linux/drivers/infiniband/hw/qib/ |
H A D | qib_ruc.c | 56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp()
|
/linux/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 262 .qp_context = p, in ipoib_cm_create_rx_qp() 589 p = wc->qp->qp_context; in ipoib_cm_handle_rx_wc() 797 struct ipoib_cm_tx *tx = wc->qp->qp_context; in ipoib_cm_handle_tx_wc() 1068 .qp_context = tx, in ipoib_cm_create_tx_qp()
|
/linux/drivers/infiniband/hw/erdma/ |
H A D | erdma_eq.c | 76 qp->ibqp.qp_context); in erdma_aeq_event_handler()
|
/linux/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-srv.c | 182 struct rtrs_srv_con *con = to_srv_con(wc->qp->qp_context); in rtrs_srv_reg_mr_done() 706 struct rtrs_srv_con *con = to_srv_con(wc->qp->qp_context); in rtrs_srv_info_rsp_done() 907 struct rtrs_srv_con *con = to_srv_con(wc->qp->qp_context); in rtrs_srv_info_req_done() 1165 struct rtrs_srv_con *con = to_srv_con(wc->qp->qp_context); in rtrs_srv_inv_rkey_done() 1222 struct rtrs_srv_con *con = to_srv_con(wc->qp->qp_context); in rtrs_srv_rdma_done()
|
H A D | rtrs-clt.c | 331 struct rtrs_clt_con *con = to_clt_con(wc->qp->qp_context); in rtrs_clt_fast_reg_done() 351 struct rtrs_clt_con *con = to_clt_con(wc->qp->qp_context); in rtrs_clt_inv_rkey_done() 603 struct rtrs_clt_con *con = to_clt_con(wc->qp->qp_context); in rtrs_clt_rdma_done() 2388 struct rtrs_clt_con *con = to_clt_con(wc->qp->qp_context); in rtrs_clt_info_req_done() 2472 struct rtrs_clt_con *con = to_clt_con(wc->qp->qp_context); in rtrs_clt_info_rsp_done()
|
H A D | rtrs.c | 266 init_attr.qp_context = con; in create_qp()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | qp.c | 791 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in hfi1_migrate_qp() 906 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in hfi1_qp_iter_cb()
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_qp.c | 1033 init_attr->qp_context = qp->ibqp.qp_context; in pvrdma_query_qp()
|
/linux/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 938 struct srpt_rdma_ch *ch = wc->qp->qp_context; in srpt_zerolength_write_done() 1391 struct srpt_rdma_ch *ch = wc->qp->qp_context; in srpt_rdma_read_done() 1756 struct srpt_rdma_ch *ch = wc->qp->qp_context; in srpt_recv_done() 1817 struct srpt_rdma_ch *ch = wc->qp->qp_context; in srpt_send_done() 1874 qp_init->qp_context = (void *)ch; in srpt_create_ch_ib()
|