Home
last modified time | relevance | path

Searched refs:to_mdev (Results 1 – 22 of 22) sorted by relevance

/linux/drivers/infiniband/hw/mthca/
H A Dmthca_provider.c59 struct mthca_dev *mdev = to_mdev(ibdev); in mthca_query_device()
139 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_port()
151 props->gid_tbl_len = to_mdev(ibdev)->limits.gid_table_len; in mthca_query_port()
153 props->pkey_tbl_len = to_mdev(ibdev)->limits.pkey_table_len; in mthca_query_port()
178 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_device()
182 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_device()
196 if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) in mthca_modify_port()
209 err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port); in mthca_modify_port()
213 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mthca_modify_port()
233 err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, in mthca_query_pkey()
[all …]
H A Dmthca_mad.c122 mthca_update_rate(to_mdev(ibdev), port_num); in smp_snoop()
123 update_sm_ah(to_mdev(ibdev), port_num, in smp_snoop()
157 mutex_lock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override()
160 mutex_unlock(&to_mdev(dev)->cap_mask_mutex); in node_desc_override()
211 forward_trap(to_mdev(ibdev), port_num, in); in mthca_process_mad()
251 err = mthca_MAD_IFC(to_mdev(ibdev), mad_flags & IB_MAD_IGNORE_MKEY, in mthca_process_mad()
257 mthca_err(to_mdev(ibdev), "MAD_IFC returned %d\n", err); in mthca_process_mad()
H A Dmthca_cq.c335 if (!mthca_is_memfree(to_mdev(cq->ibcq.device)) && in mthca_cq_resize_copy_cqes()
657 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_poll_cq()
730 mthca_write64(dbhi, 0xffffffff, to_mdev(cq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_tavor_arm_cq()
731 MTHCA_GET_DOORBELL_LOCK(&to_mdev(cq->device)->doorbell_lock)); in mthca_tavor_arm_cq()
762 to_mdev(ibcq->device)->kar + MTHCA_CQ_DOORBELL, in mthca_arbel_arm_cq()
763 MTHCA_GET_DOORBELL_LOCK(&to_mdev(ibcq->device)->doorbell_lock)); in mthca_arbel_arm_cq()
H A Dmthca_srq.c375 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_modify_srq()
398 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_query_srq()
486 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_tavor_post_srq_recv()
580 struct mthca_dev *dev = to_mdev(ibsrq->device); in mthca_arbel_post_srq_recv()
H A Dmthca_mcg.c122 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_attach()
216 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_detach()
H A Dmthca_qp.c433 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_query_qp()
562 struct mthca_dev *dev = to_mdev(ibqp->device); in __mthca_modify_qp()
861 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_modify_qp()
1628 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_tavor_post_send()
1826 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_tavor_post_receive()
1931 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_arbel_post_send()
2165 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_arbel_post_receive()
H A Dmthca_av.c295 struct mthca_dev *dev = to_mdev(ibah->device); in mthca_ah_query()
H A Dmthca_dev.h575 static inline struct mthca_dev *to_mdev(struct ib_device *ibdev) in to_mdev() function
/linux/drivers/infiniband/hw/mlx5/
H A Dib_virt.c53 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_get_vf_config()
95 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_set_vf_link_state()
129 dev = to_mdev(device); in mlx5_ib_get_vf_stats()
154 struct mlx5_ib_dev *dev = to_mdev(device); in set_vf_node_guid()
178 struct mlx5_ib_dev *dev = to_mdev(device); in set_vf_port_guid()
214 struct mlx5_ib_dev *dev = to_mdev(device); in mlx5_ib_get_vf_guid()
H A Dgsi.c91 struct mlx5_ib_dev *dev = to_mdev(pd->device); in mlx5_ib_create_gsi()
171 struct mlx5_ib_dev *dev = to_mdev(mqp->ibqp.device); in mlx5_ib_destroy_gsi()
223 struct mlx5_ib_dev *dev = to_mdev(qp->device); in modify_to_rts()
263 struct mlx5_ib_dev *dev = to_mdev(device); in setup_qp()
325 struct mlx5_ib_dev *dev = to_mdev(qp->device); in mlx5_ib_gsi_modify_qp()
365 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in mlx5_ib_add_outstanding_wr()
415 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); in get_tx_qp()
H A Dcounters.c210 mlx5_fc_destroy(to_mdev(counters->device)->mdev, in mlx5_ib_destroy_counters()
281 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_alloc_hw_device_stats()
290 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_alloc_hw_port_stats()
391 struct mlx5_ib_dev *dev = to_mdev(ibdev); in do_get_hw_stats()
462 struct mlx5_ib_dev *dev = to_mdev(counter->device); in do_per_qp_get_op_stat()
499 struct mlx5_ib_dev *dev = to_mdev(ibdev); in do_get_op_stat()
533 struct mlx5_ib_dev *dev = to_mdev(ibdev); in do_get_op_stats()
555 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_get_hw_stats()
584 struct mlx5_ib_dev *dev = to_mdev(counter->device); in mlx5_ib_counter_alloc_stats()
592 struct mlx5_ib_dev *dev = to_mdev(counter->device); in mlx5_ib_counter_update_stats()
[all …]
H A Dmacsec.c219 struct mlx5_ib_dev *dev = to_mdev(attr->device); in mlx5r_add_gid_macsec_operations()
262 ret = set_roce_addr(to_mdev(physical_gid->device), in mlx5r_add_gid_macsec_operations()
289 set_roce_addr(to_mdev(physical_gid->device), physical_gid->port_num, in mlx5r_add_gid_macsec_operations()
303 struct mlx5_ib_dev *dev = to_mdev(attr->device); in mlx5r_del_gid_macsec_operations()
345 set_roce_addr(to_mdev(physical_gid->device), in mlx5r_del_gid_macsec_operations()
H A Dmad.c305 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_ib_process_mad()
346 err = mlx5_MAD_IFC(to_mdev(ibdev), mad_flags & IB_MAD_IGNORE_MKEY, in mlx5_ib_process_mad()
405 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, 1, NULL, NULL, in_mad, in mlx5_query_mad_ifc_smp_attr_node_info()
544 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_pkey()
573 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids()
584 err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, in mlx5_query_mad_ifc_gids()
600 struct mlx5_ib_dev *dev = to_mdev(ibdev); in mlx5_query_mad_ifc_port()
H A Dah.c105 struct mlx5_ib_dev *dev = to_mdev(ibah->device); in mlx5_ib_create_ah()
H A Dwr.c642 struct mlx5_ib_dev *dev = to_mdev(pd->ibpd.device); in set_reg_wr()
654 to_mdev(qp->ibqp.device), in set_reg_wr()
660 mlx5_ib_warn(to_mdev(qp->ibqp.device), in set_reg_wr()
1055 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_post_send()
1212 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_post_recv()
/linux/drivers/dma/
H A Daltera-msgdma.c201 #define to_mdev(chan) container_of(chan, struct msgdma_device, dmachan) macro
306 struct msgdma_device *mdev = to_mdev(tx->chan); in msgdma_tx_submit()
335 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_memcpy()
393 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_slave_sg()
462 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_dma_config()
569 struct msgdma_device *mdev = to_mdev(chan); in msgdma_issue_pending()
639 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_free_chan_resources()
656 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_alloc_chan_resources()
/linux/drivers/infiniband/hw/mlx4/
H A Dcm.c156 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in id_map_find_by_sl_id()
201 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in sl_id_map_add()
236 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_alloc()
245 ent->dev = to_mdev(ibdev); in id_map_alloc()
268 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_get()
284 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in schedule_delayed()
429 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in mlx4_ib_demux_cm_handler()
H A Dmad.c230 struct mlx4_ib_dev *dev = to_mdev(ibdev); in smp_snoop()
376 spin_lock_irqsave(&to_mdev(dev)->sm_lock, flags); in node_desc_override()
379 spin_unlock_irqrestore(&to_mdev(dev)->sm_lock, flags); in node_desc_override()
435 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_find_real_gid()
686 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_demux_mad()
833 forward_trap(to_mdev(ibdev), port_num, in_mad); in ib_process_mad()
866 err = mlx4_MAD_IFC(to_mdev(ibdev), in ib_process_mad()
877 if (!mlx4_is_slave(to_mdev(ibdev)->dev)) in ib_process_mad()
942 struct mlx4_ib_dev *dev = to_mdev(ibdev); in iboe_process_mad()
989 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_process_mad()
[all …]
H A Dah.c46 struct mlx4_dev *dev = to_mdev(ib_ah->device)->dev; in create_ib_ah()
79 struct mlx4_ib_dev *ibdev = to_mdev(ib_ah->device); in create_iboe_ah()
H A Dmcg.c885 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_mcg_demux_handler()
934 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_mcg_multiplex_handler()
H A Dalias_GUID.c489 struct mlx4_ib_dev *dev = to_mdev(ibdev); in set_guid_rec()
/linux/drivers/most/
H A Dmost_usb.c124 #define to_mdev(d) container_of(d, struct most_dev, iface) macro
238 struct most_dev *mdev = to_mdev(iface); in hdm_poison_channel()
343 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_write_completion()
395 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_read_completion()
460 struct most_dev *mdev = to_mdev(iface); in hdm_enqueue()
537 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_dma_alloc()
545 struct most_dev *mdev = to_mdev(mbo->ifp); in hdm_dma_free()
571 struct most_dev *mdev = to_mdev(iface); in hdm_configure_channel()
651 struct most_dev *mdev = to_mdev(iface); in hdm_request_netinfo()