| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_provider.c | 59 struct mthca_dev *mdev = to_mdev(ibdev); in mthca_query_device() 139 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_port() 151 props->gid_tbl_len = to_mdev(ibdev)->limits.gid_table_len; in mthca_query_port() 153 props->pkey_tbl_len = to_mdev(ibdev)->limits.pkey_table_len; in mthca_query_port() 178 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_device() 182 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_device() 196 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_port() 209 err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port); in mthca_modify_port() 213 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_port() 233 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_pkey() [all …]
|
| H A D | mthca_mad.c | 122 mthca_update_rate(to_mdev(ibdev), port_num); in smp_snoop() 123 update_sm_ah(to_mdev(ibdev), port_num, in smp_snoop() 157 mutex_lock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override() 160 mutex_unlock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override() 211 forward_trap(to_mdev(ibdev), port_num, in); in mthca_process_mad() 251 err = mthca_MAD_IFC(to_mdev(ibdev), mad_flags & IB_MAD_IGNORE_MKEY, in mthca_process_mad() 257 mthca_err(to_mdev(ibdev), "MAD_IFC returned %d\n", err); in mthca_process_mad()
|
| H A D | mthca_cq.c | 335 if (!mthca_is_memfree(to_mdev(cq->ibcq.device)) && in mthca_cq_resize_copy_cqes() 657 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_poll_cq() 730 mthca_write64(dbhi, 0xffffffff, to_mdev(cq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_tavor_arm_cq() 731 MTHCA_GET_DOORBELL_LOCK(&to_mdev(cq->device)->doorbell_lock)); in mthca_tavor_arm_cq() 762 to_mdev(ibcq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_arbel_arm_cq() 763 MTHCA_GET_DOORBELL_LOCK(&to_mdev(ibcq->device)->doorbell_lock)); in mthca_arbel_arm_cq()
|
| H A D | mthca_srq.c | 375 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_modify_srq() 398 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_query_srq() 486 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_tavor_post_srq_recv() 580 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_arbel_post_srq_recv()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | mr.c | 67 err = mlx4_mr_alloc(to_mdev(pd->device)->dev, to_mpd(pd)->pdn, 0, in mlx4_ib_get_dma_mr() 72 err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr() 82 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr() 145 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_reg_user_mr() 189 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_reg_user_mr() 205 struct mlx4_ib_dev *dev = to_mdev(mr->device); in mlx4_ib_rereg_user_mr() 345 ret = mlx4_mr_free(to_mdev(ibmr->device)->dev, &mr->mmr); in mlx4_ib_dereg_mr() 357 struct mlx4_ib_dev *dev = to_mdev(ibmw->device); in mlx4_ib_alloc_mw() 382 mlx4_mw_free(to_mdev(ibmw->device)->dev, &mw->mmw); in mlx4_ib_dealloc_mw() 389 struct mlx4_ib_dev *dev = to_mdev(pd->device); in mlx4_ib_alloc_mr()
|
| H A D | main.c | 129 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_get_netdev() 257 struct mlx4_ib_dev *ibdev = to_mdev(attr->device); in mlx4_ib_add_gid() 349 struct mlx4_ib_dev *ibdev = to_mdev(attr->device); in mlx4_ib_del_gid() 438 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_query_device() 473 err = mlx4_MAD_IFC(to_mdev(ibdev), MLX4_MAD_IFC_IGNORE_KEYS, in mlx4_ib_query_device() 648 struct mlx4_dev *dev = to_mdev(device)->dev; in mlx4_ib_port_link_layer() 672 if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) in ib_link_query_port() 675 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, in ib_link_query_port() 691 props->gid_tbl_len = to_mdev(ibdev)->dev->caps.gid_table_len[port]; in ib_link_query_port() 692 props->max_msg_sz = to_mdev(ibdev)->dev->caps.max_msg_sz; in ib_link_query_port() [all …]
|
| H A D | srq.c | 76 struct mlx4_ib_dev *dev = to_mdev(ib_srq->device); in mlx4_ib_create_srq() 227 struct mlx4_ib_dev *dev = to_mdev(ibsrq->device); in mlx4_ib_modify_srq() 252 struct mlx4_ib_dev *dev = to_mdev(ibsrq->device); in mlx4_ib_query_srq() 270 struct mlx4_ib_dev *dev = to_mdev(srq->device); in mlx4_ib_destroy_srq() 317 struct mlx4_ib_dev *mdev = to_mdev(ibsrq->device); in mlx4_ib_post_srq_recv()
|
| H A D | cm.c | 156 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in id_map_find_by_sl_id() 201 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in sl_id_map_add() 236 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_alloc() 245 ent->dev = to_mdev(ibdev); in id_map_alloc() 268 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_get() 284 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in schedule_delayed() 429 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in mlx4_ib_demux_cm_handler()
|
| H A D | cq.c | 96 struct mlx4_ib_dev *dev = to_mdev(cq->device); in mlx4_ib_modify_cq() 185 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_create_cq() 386 struct mlx4_ib_dev *dev = to_mdev(ibcq->device); in mlx4_ib_resize_cq() 484 struct mlx4_ib_dev *dev = to_mdev(cq->device); in mlx4_ib_destroy_cq() 700 struct mlx4_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx4_ib_poll_one() 720 mqp = __mlx4_qp_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 732 msrq = mlx4_srq_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 842 if (mlx4_is_mfunc(to_mdev(cq->ibcq.device)->dev)) { in mlx4_ib_poll_one() 888 struct mlx4_ib_dev *mdev = to_mdev(cq->ibcq.device); in mlx4_ib_poll_cq() 914 to_mdev(ibcq->device)->uar_map, in mlx4_ib_arm_cq() [all …]
|
| H A D | mad.c | 230 struct mlx4_ib_dev *dev = to_mdev(ibdev); in smp_snoop() 376 spin_lock_irqsave(&to_mdev(dev)->sm_lock, flags); in node_desc_override() 379 spin_unlock_irqrestore(&to_mdev(dev)->sm_lock, flags); in node_desc_override() 435 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_find_real_gid() 686 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_demux_mad() 833 forward_trap(to_mdev(ibdev), port_num, in_mad); in ib_process_mad() 866 err = mlx4_MAD_IFC(to_mdev(ibdev), in ib_process_mad() 877 if (!mlx4_is_slave(to_mdev(ibdev)->dev)) in ib_process_mad() 942 struct mlx4_ib_dev *dev = to_mdev(ibdev); in iboe_process_mad() 989 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_process_mad() [all …]
|
| H A D | qp.c | 768 err = create_qp_rss(to_mdev(pd->device), init_attr, &ucmd, qp); in _mlx4_ib_create_qp_rss() 784 struct mlx4_ib_dev *dev = to_mdev(context->ibucontext.device); in mlx4_ib_alloc_wqn() 834 struct mlx4_ib_dev *dev = to_mdev(context->ibucontext.device); in mlx4_ib_release_wqn() 862 struct mlx4_ib_dev *dev = to_mdev(pd->device); in create_rq() 997 struct mlx4_ib_dev *dev = to_mdev(pd->device); in create_qp_common() 1572 if (!(to_mdev(pd->device)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC)) in _mlx4_ib_create_qp() 1595 int res = mlx4_qp_reserve_range(to_mdev(pd->device)->dev, in _mlx4_ib_create_qp() 1602 sqpn = get_sqp_num(to_mdev(pd->device), init_attr); in _mlx4_ib_create_qp() 1632 struct mlx4_ib_dev *dev = to_mdev(device); in mlx4_ib_create_qp() 1669 struct mlx4_ib_dev *dev = to_mdev(qp->device); in _mlx4_ib_destroy_qp() [all …]
|
| H A D | ah.c | 46 struct mlx4_dev *dev = to_mdev(ib_ah->device)->dev; in create_ib_ah() 79 struct mlx4_ib_dev *ibdev = to_mdev(ib_ah->device); in create_iboe_ah()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | ib_virt.c | 53 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_get_vf_config() 95 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_set_vf_link_state() 129 dev = to_mdev(device); in mlx5_ib_get_vf_stats() 154 struct mlx5_ib_dev *dev = to_mdev(device); in set_vf_node_guid() 178 struct mlx5_ib_dev *dev = to_mdev(device); in set_vf_port_guid() 214 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_get_vf_guid()
|
| H A D | gsi.c | 91 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_create_gsi() 171 struct mlx5_ib_dev *dev = to_mdev(mqp->ibqp.device); in mlx5_ib_destroy_gsi() 223 struct mlx5_ib_dev *dev = to_mdev(qp->device); in modify_to_rts() 263 struct mlx5_ib_dev *dev = to_mdev(device); in setup_qp() 325 struct mlx5_ib_dev *dev = to_mdev(qp->device); in mlx5_ib_gsi_modify_qp() 365 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in mlx5_ib_add_outstanding_wr() 415 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in get_tx_qp()
|
| H A D | counters.c | 210 mlx5_fc_destroy(to_mdev(counters->device)->mdev, in mlx5_ib_destroy_counters() 281 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_alloc_hw_device_stats() 290 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_alloc_hw_port_stats() 391 struct mlx5_ib_dev *dev = to_mdev(ibdev); in do_get_hw_stats() 462 struct mlx5_ib_dev *dev = to_mdev(counter->device); in do_per_qp_get_op_stat() 499 struct mlx5_ib_dev *dev = to_mdev(ibdev); in do_get_op_stat() 533 struct mlx5_ib_dev *dev = to_mdev(ibdev); in do_get_op_stats() 555 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_get_hw_stats() 584 struct mlx5_ib_dev *dev = to_mdev(counter->device); in mlx5_ib_counter_alloc_stats() 592 struct mlx5_ib_dev *dev = to_mdev(counter->device); in mlx5_ib_counter_update_stats() [all …]
|
| H A D | macsec.c | 219 struct mlx5_ib_dev *dev = to_mdev(attr->device); in mlx5r_add_gid_macsec_operations() 262 ret = set_roce_addr(to_mdev(physical_gid->device), in mlx5r_add_gid_macsec_operations() 289 set_roce_addr(to_mdev(physical_gid->device), physical_gid->port_num, in mlx5r_add_gid_macsec_operations() 303 struct mlx5_ib_dev *dev = to_mdev(attr->device); in mlx5r_del_gid_macsec_operations() 345 set_roce_addr(to_mdev(physical_gid->device), in mlx5r_del_gid_macsec_operations()
|
| H A D | srq.c | 47 struct mlx5_ib_dev *dev = to_mdev(pd->device); in create_srq_user() 195 struct mlx5_ib_dev *dev = to_mdev(ib_srq->device); in mlx5_ib_create_srq() 331 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_modify_srq() 356 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_query_srq() 380 struct mlx5_ib_dev *dev = to_mdev(srq->device); in mlx5_ib_destroy_srq() 415 struct mlx5_ib_dev *dev = to_mdev(ibsrq->device); in mlx5_ib_post_srq_recv()
|
| H A D | mad.c | 305 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_process_mad() 346 err = mlx5_MAD_IFC(to_mdev(ibdev), mad_flags & IB_MAD_IGNORE_MKEY, in mlx5_ib_process_mad() 405 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, 1, NULL, NULL, in_mad, in mlx5_query_mad_ifc_smp_attr_node_info() 544 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_pkey() 573 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids() 584 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids() 600 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_mad_ifc_port()
|
| H A D | mr.c | 67 struct mlx5_ib_dev *dev = to_mdev(pd->device); in set_mkc_access_pd_addr_fields() 1054 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_get_dma_mr() 1145 struct mlx5_ib_dev *dev = to_mdev(pd->device); in alloc_cacheable_mr() 1198 struct mlx5_ib_dev *dev = to_mdev(pd->device); in reg_create_crossing_vhca_mr() 1260 struct mlx5_ib_dev *dev = to_mdev(pd->device); in reg_create() 1365 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_get_dm_mr() 1429 struct mlx5_core_dev *dev = to_mdev(dm->device)->mdev; in mlx5_ib_reg_dm_mr() 1462 struct mlx5_ib_dev *dev = to_mdev(pd->device); in create_real_mr() 1520 struct mlx5_ib_dev *dev = to_mdev(pd->device); in create_user_odp_mr() 1582 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_reg_user_mr() [all …]
|
| H A D | qos.c | 17 struct mlx5_ib_dev *dev = to_mdev(device); in pp_is_supported() 47 dev = to_mdev(c->ibucontext.device); in UVERBS_HANDLER()
|
| H A D | cq.c | 54 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event() 173 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.device); in handle_responder() 454 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one() 587 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in poll_soft_wc() 614 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_poll_cq() 649 struct mlx5_core_dev *mdev = to_mdev(ibcq->device)->mdev; in mlx5_ib_arm_cq() 959 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_create_cq() 1063 struct mlx5_ib_dev *dev = to_mdev(cq->device); in mlx5_ib_pre_destroy_cq() 1071 destroy_cq_kernel(to_mdev(cq->device), to_mcq(cq)); in mlx5_ib_post_destroy_cq() 1157 struct mlx5_ib_dev *dev = to_mdev(cq->device); in mlx5_ib_modify_cq() [all …]
|
| H A D | dm.c | 167 struct mlx5_ib_dev *dev = to_mdev(dm->base.ibdm.device); in mlx5_ib_dm_memic_free() 209 struct mlx5_ib_dev *dev = to_mdev(uobj->context->device); in UVERBS_HANDLER() 278 struct mlx5_dm *dm_db = &to_mdev(ctx->device)->dm; in handle_alloc_dm_memic() 357 struct mlx5_core_dev *dev = to_mdev(ctx->device)->mdev; in handle_alloc_dm_sw_icm() 435 mlx5_ib_dbg(to_mdev(ibdev), "alloc_dm req: dm_type=%d user_length=0x%llx log_alignment=%d\n", in mlx5_ib_alloc_dm() 474 struct mlx5_core_dev *dev = to_mdev(dm->base.ibdm.device)->mdev; in mlx5_dm_icm_dealloc()
|
| H A D | main.c | 112 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_port_link_layer() 538 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_query_port_roce() 689 return set_roce_addr(to_mdev(attr->device), attr->port_num, in mlx5_ib_add_gid() 698 ret = set_roce_addr(to_mdev(attr->device), attr->port_num, in mlx5_ib_del_gid() 731 if (mlx5_use_mad_ifc(to_mdev(ibdev))) in mlx5_get_vport_access_method() 774 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_system_image_guid() 806 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_max_pkeys() 827 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_vendor_id() 949 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_query_device() 1369 struct mlx5_ib_dev *dev = to_mdev(ibdev); in translate_active_width() [all …]
|
| /linux/drivers/dma/ |
| H A D | altera-msgdma.c | 201 #define to_mdev(chan) container_of(chan, struct msgdma_device, dmachan) macro 306 struct msgdma_device *mdev = to_mdev(tx->chan); in msgdma_tx_submit() 335 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_memcpy() 393 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_slave_sg() 462 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_dma_config() 569 struct msgdma_device *mdev = to_mdev(chan); in msgdma_issue_pending() 639 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_free_chan_resources() 656 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_alloc_chan_resources()
|
| /linux/drivers/most/ |
| H A D | most_usb.c | 124 #define to_mdev(d) container_of(d, struct most_dev, iface) macro 238 struct most_dev *mdev = to_mdev(iface); in hdm_poison_channel() 343 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_write_completion() 395 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_read_completion() 460 struct most_dev *mdev = to_mdev(iface); in hdm_enqueue() 537 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_dma_alloc() 545 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_dma_free() 571 struct most_dev *mdev = to_mdev(iface); in hdm_configure_channel() 651 struct most_dev *mdev = to_mdev(iface); in hdm_request_netinfo()
|