/freebsd/sys/dev/mthca/ |
H A D | mthca_provider.c | 66 struct mthca_dev *mdev = to_mdev(ibdev); in mthca_query_device() 155 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_port() 167 props->gid_tbl_len = to_mdev(ibdev)->limits.gid_table_len; in mthca_query_port() 169 props->pkey_tbl_len = to_mdev(ibdev)->limits.pkey_table_len; in mthca_query_port() 194 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_device() 198 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_device() 212 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_port() 225 err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port); in mthca_modify_port() 229 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_port() 249 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_pkey() [all …]
|
H A D | mthca_mad.c | 120 mthca_update_rate(to_mdev(ibdev), port_num); in smp_snoop() 121 update_sm_ah(to_mdev(ibdev), port_num, in smp_snoop() 155 mutex_lock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override() 158 mutex_unlock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override() 220 forward_trap(to_mdev(ibdev), port_num, in_mad); in mthca_process_mad() 260 err = mthca_MAD_IFC(to_mdev(ibdev), in mthca_process_mad() 267 mthca_err(to_mdev(ibdev), "MAD_IFC returned %d\n", err); in mthca_process_mad()
|
H A D | mthca_cq.c | 340 if (!mthca_is_memfree(to_mdev(cq->ibcq.device)) && in mthca_cq_resize_copy_cqes() 662 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_poll_cq() 735 mthca_write64(dbhi, 0xffffffff, to_mdev(cq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_tavor_arm_cq() 736 MTHCA_GET_DOORBELL_LOCK(&to_mdev(cq->device)->doorbell_lock)); in mthca_tavor_arm_cq() 767 to_mdev(ibcq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_arbel_arm_cq() 768 MTHCA_GET_DOORBELL_LOCK(&to_mdev(ibcq->device)->doorbell_lock)); in mthca_arbel_arm_cq()
|
H A D | mthca_srq.c | 369 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_modify_srq() 392 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_query_srq() 480 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_tavor_post_srq_recv() 580 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_arbel_post_srq_recv()
|
H A D | mthca_mcg.c | 122 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_attach() 216 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_detach()
|
/freebsd/sys/dev/mlx4/mlx4_ib/ |
H A D | mlx4_ib_mr.c | 70 err = mlx4_mr_alloc(to_mdev(pd->device)->dev, to_mpd(pd)->pdn, 0, in mlx4_ib_get_dma_mr() 75 err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr() 85 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr() 141 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_reg_user_mr() 181 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_reg_user_mr() 197 struct mlx4_ib_dev *dev = to_mdev(mr->device); in mlx4_ib_rereg_user_mr() 333 ret = mlx4_mr_free(to_mdev(ibmr->device)->dev, &mr->mmr); in mlx4_ib_dereg_mr() 346 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_alloc_mw() 380 mlx4_mw_free(to_mdev(ibmw->device)->dev, &mw->mmw); in mlx4_ib_dealloc_mw() 389 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_alloc_mr() [all …]
|
H A D | mlx4_ib_main.c | 136 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_get_netdev() 262 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_add_gid() 340 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_del_gid() 442 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_query_device() 476 err = mlx4_MAD_IFC(to_mdev(ibdev), MLX4_MAD_IFC_IGNORE_KEYS, in mlx4_ib_query_device() 586 struct mlx4_dev *dev = to_mdev(device)->dev; in mlx4_ib_port_link_layer() 610 if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) in ib_link_query_port() 613 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, in ib_link_query_port() 629 props->gid_tbl_len = to_mdev(ibdev)->dev->caps.gid_table_len[port]; in ib_link_query_port() 630 props->max_msg_sz = to_mdev(ibdev)->dev->caps.max_msg_sz; in ib_link_query_port() [all …]
|
H A D | mlx4_ib_cm.c | 146 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in id_map_find_by_sl_id() 194 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_find_del() 212 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in sl_id_map_add() 247 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_alloc() 258 ent->dev = to_mdev(ibdev); in id_map_alloc() 262 spin_lock(&to_mdev(ibdev)->sriov.id_map_lock); in id_map_alloc() 287 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_get() 303 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in schedule_delayed()
|
H A D | mlx4_ib_srq.c | 76 struct mlx4_ib_dev *dev = to_mdev(ib_srq->device); in mlx4_ib_create_srq() 227 struct mlx4_ib_dev *dev = to_mdev(ibsrq->device); in mlx4_ib_modify_srq() 252 struct mlx4_ib_dev *dev = to_mdev(ibsrq->device); in mlx4_ib_query_srq() 270 struct mlx4_ib_dev *dev = to_mdev(srq->device); in mlx4_ib_destroy_srq() 316 struct mlx4_ib_dev *mdev = to_mdev(ibsrq->device); in mlx4_ib_post_srq_recv()
|
H A D | mlx4_ib_cq.c | 97 struct mlx4_ib_dev *dev = to_mdev(cq->device); in mlx4_ib_modify_cq() 180 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_create_cq() 370 struct mlx4_ib_dev *dev = to_mdev(ibcq->device); in mlx4_ib_resize_cq() 468 struct mlx4_ib_dev *dev = to_mdev(cq->device); in mlx4_ib_destroy_cq() 688 struct mlx4_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx4_ib_poll_one() 708 mqp = __mlx4_qp_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 720 msrq = mlx4_srq_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 828 if (mlx4_is_mfunc(to_mdev(cq->ibcq.device)->dev)) { in mlx4_ib_poll_one() 873 struct mlx4_ib_dev *mdev = to_mdev(cq->ibcq.device); in mlx4_ib_poll_cq() 898 struct mlx4_ib_dev *mdev = to_mdev(cq->ibcq.device); in mlx4_ib_arm_cq()
|
H A D | mlx4_ib_mad.c | 227 struct mlx4_ib_dev *dev = to_mdev(ibdev); in smp_snoop() 372 spin_lock_irqsave(&to_mdev(dev)->sm_lock, flags); in node_desc_override() 375 spin_unlock_irqrestore(&to_mdev(dev)->sm_lock, flags); in node_desc_override() 430 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_find_real_gid() 643 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_demux_mad() 794 forward_trap(to_mdev(ibdev), port_num, in_mad); in ib_process_mad() 827 err = mlx4_MAD_IFC(to_mdev(ibdev), in ib_process_mad() 838 if (!mlx4_is_slave(to_mdev(ibdev)->dev)) in ib_process_mad() 904 struct mlx4_ib_dev *dev = to_mdev(ibdev); in iboe_process_mad() 953 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_process_mad() [all …]
|
H A D | mlx4_ib_qp.c | 1186 if (!(to_mdev(pd->device)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC)) in _mlx4_ib_create_qp() 1201 err = create_qp_common(to_mdev(pd->device), pd, init_attr, in _mlx4_ib_create_qp() 1222 int res = mlx4_qp_reserve_range(to_mdev(pd->device)->dev, 1, 1, &sqpn, 0); in _mlx4_ib_create_qp() 1227 sqpn = get_sqp_num(to_mdev(pd->device), init_attr); in _mlx4_ib_create_qp() 1230 err = create_qp_common(to_mdev(pd->device), pd, init_attr, udata, in _mlx4_ib_create_qp() 1254 struct mlx4_ib_dev *dev = to_mdev(device); in mlx4_ib_create_qp() 1285 struct mlx4_ib_dev *dev = to_mdev(qp->device); in _mlx4_ib_destroy_qp() 1630 struct mlx4_ib_dev *dev = to_mdev(ibqp->device); in __mlx4_ib_modify_qp() 1774 u8 port_num = mlx4_is_bonded(to_mdev(ibqp->device)->dev) ? 1 : in __mlx4_ib_modify_qp() 2170 struct mlx4_ib_dev *dev = to_mdev(ibqp->device); in _mlx4_ib_modify_qp() [all …]
|
H A D | mlx4_ib_ah.c | 49 struct mlx4_dev *dev = to_mdev(ib_ah->device)->dev; in create_ib_ah() 77 struct mlx4_ib_dev *ibdev = to_mdev(ib_ah->device); in create_iboe_ah()
|
/freebsd/sys/dev/mlx5/mlx5_ib/ |
H A D | mlx5_ib_mad.c | 91 err = mlx5_MAD_IFC(to_mdev(ibdev), in process_mad() 184 struct mlx5_ib_dev *dev = to_mdev(ibdev); in process_pma_cmd() 238 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_process_mad() 301 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, 1, NULL, NULL, in_mad, in mlx5_query_mad_ifc_smp_attr_node_info() 440 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_pkey() 469 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids() 480 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids() 496 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_mad_ifc_port()
|
H A D | mlx5_ib_gsi.c | 114 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_gsi_create_qp() 211 struct mlx5_ib_dev *dev = to_mdev(qp->device); in mlx5_ib_gsi_destroy_qp() 271 struct mlx5_ib_dev *dev = to_mdev(qp->device); in modify_to_rts() 311 struct mlx5_ib_dev *dev = to_mdev(device); in setup_qp() 372 struct mlx5_ib_dev *dev = to_mdev(qp->device); in mlx5_ib_gsi_modify_qp() 413 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in mlx5_ib_add_outstanding_wr() 464 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in get_tx_qp()
|
H A D | mlx5_ib_srq.c | 73 struct mlx5_ib_dev *dev = to_mdev(pd->device); in create_srq_user() 242 struct mlx5_ib_dev *dev = to_mdev(ib_srq->device); in mlx5_ib_create_srq() 348 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_modify_srq() 373 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_query_srq() 397 struct mlx5_ib_dev *dev = to_mdev(srq->device); in mlx5_ib_destroy_srq() 435 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_post_srq_recv()
|
H A D | mlx5_ib_main.c | 82 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_port_link_layer() 144 struct mlx5_ib_dev *ibdev = to_mdev(device); in mlx5_ib_get_netdev() 288 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_query_port_roce() 403 struct mlx5_ib_dev *dev = to_mdev(device); in set_roce_addr() 489 if (mlx5_use_mad_ifc(to_mdev(ibdev))) in mlx5_get_vport_access_method() 524 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_system_image_guid() 556 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_max_pkeys() 577 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_vendor_id() 642 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_query_device() 842 struct mlx5_ib_dev *dev = to_mdev(ibde in translate_active_width() [all...] |
H A D | mlx5_ib_mr.c | 479 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_get_dma_mr() 581 struct mlx5_ib_dev *dev = to_mdev(pd->device); in prep_umr_wqe_common() 628 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mr_umem_get() 669 struct mlx5_ib_dev *dev = to_mdev(pd->device); in reg_umr() 875 struct mlx5_ib_dev *dev = to_mdev(pd->device); in reg_create() 961 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_reg_user_mr() 1061 struct mlx5_ib_dev *dev = to_mdev(pd->device); in rereg_umr() 1129 struct mlx5_ib_dev *dev = to_mdev(ib_mr->device); in mlx5_ib_rereg_user_mr() 1266 struct mlx5_ib_dev *dev = to_mdev(mr->ibmr.device); in clean_mr() 1309 struct mlx5_ib_dev *dev = to_mdev(ibmr->device); in mlx5_ib_dereg_mr() [all …]
|
H A D | mlx5_ib_cq.c | 46 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event() 168 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.device); in handle_responder() 518 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one() 649 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in poll_soft_wc() 672 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_poll_cq() 702 struct mlx5_core_dev *mdev = to_mdev(ibcq->device)->mdev; in mlx5_ib_arm_cq() 922 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_create_cq() 1019 struct mlx5_ib_dev *dev = to_mdev(cq->device); in mlx5_ib_destroy_cq() 1097 struct mlx5_ib_dev *dev = to_mdev(cq->device); in mlx5_ib_modify_cq() 1185 struct mlx5_ib_dev *dev = to_mdev(c in copy_resize_cqes() [all...] |
H A D | mlx5_ib_devx.c | 994 dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1038 dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1071 dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1402 struct mlx5_ib_dev *dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1505 struct mlx5_ib_dev *mdev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1553 struct mlx5_ib_dev *mdev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1630 struct mlx5_ib_dev *dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1690 struct mlx5_ib_dev *mdev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 1913 struct mlx5_ib_dev *dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() 2170 struct mlx5_ib_dev *dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER() [all …]
|
H A D | mlx5_ib_qp.c | 137 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_read_user_wqe() 2104 dev = to_mdev(pd->device); in mlx5_ib_create_qp() 2123 dev = to_mdev(to_mxrcd(init_attr->xrcd)->ibxrcd.device); in mlx5_ib_create_qp() 2192 struct mlx5_ib_dev *dev = to_mdev(qp->device); in mlx5_ib_destroy_qp() 2212 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.device); in to_mlx5_access_flags() 2718 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in __mlx5_ib_modify_qp() 2960 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_modify_qp() 3771 mlx5_ib_warn(to_mdev(qp->ibqp.device), in set_reg_wr() 3900 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_post_send() 4217 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_post_recv() [all …]
|
H A D | mlx5_ib_ah.c | 66 struct mlx5_ib_dev *dev = to_mdev(ibah->device); in mlx5_ib_create_ah()
|
/freebsd/contrib/ofed/libmlx4/ |
H A D | mlx4.c | 153 struct mlx4_device *dev = to_mdev(&v_device->device); in mlx4_init_context() 287 munmap(context->uar, to_mdev(&v_device->device)->page_size); in mlx4_uninit_context() 289 munmap(context->bf_page, to_mdev(&v_device->device)->page_size); in mlx4_uninit_context() 292 to_mdev(&v_device->device)->page_size); in mlx4_uninit_context()
|
H A D | dbrec.c | 58 int ps = to_mdev(context->ibv_ctx.device)->page_size; in __add_page() 122 uintptr_t ps = to_mdev(context->ibv_ctx.device)->page_size; in mlx4_free_db()
|
/freebsd/contrib/ofed/libmlx5/ |
H A D | dbrec.c | 52 int ps = to_mdev(context->ibv_ctx.device)->page_size; in __add_page() 119 uintptr_t ps = to_mdev(context->ibv_ctx.device)->page_size; in mlx5_free_db()
|