Home
last modified time | relevance | path

Searched refs:ib_dev (Results 1 – 25 of 85) sorted by relevance

1234

/linux/drivers/target/
H A Dtarget_core_iblock.c60 struct iblock_dev *ib_dev = NULL; in iblock_alloc_device() local
62 ib_dev = kzalloc_obj(struct iblock_dev); in iblock_alloc_device()
63 if (!ib_dev) { in iblock_alloc_device()
67 ib_dev->ibd_exclusive = true; in iblock_alloc_device()
69 ib_dev->ibd_plug = kzalloc_objs(*ib_dev->ibd_plug, nr_cpu_ids); in iblock_alloc_device()
70 if (!ib_dev->ibd_plug) in iblock_alloc_device()
75 return &ib_dev->dev; in iblock_alloc_device()
78 kfree(ib_dev); in iblock_alloc_device()
84 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); in iblock_configure_unmap() local
87 ib_dev->ibd_bd); in iblock_configure_unmap()
[all …]
/linux/drivers/infiniband/core/
H A Droce_gid_mgmt.c82 unsigned long roce_gid_type_mask_support(struct ib_device *ib_dev, u32 port) in roce_gid_type_mask_support() argument
87 if (!rdma_protocol_roce(ib_dev, port)) in roce_gid_type_mask_support()
91 if (PORT_CAP_TO_GID_TYPE[i].is_supported(ib_dev, port)) in roce_gid_type_mask_support()
98 static void update_gid(enum gid_op_type gid_op, struct ib_device *ib_dev, in update_gid() argument
103 unsigned long gid_type_mask = roce_gid_type_mask_support(ib_dev, port); in update_gid()
110 ib_cache_gid_add(ib_dev, port, in update_gid()
114 ib_cache_gid_del(ib_dev, port, in update_gid()
147 is_eth_port_of_netdev_filter(struct ib_device *ib_dev, u32 port, in is_eth_port_of_netdev_filter() argument
171 is_eth_port_inactive_slave_filter(struct ib_device *ib_dev, u32 port, in is_eth_port_inactive_slave_filter() argument
201 is_ndev_for_default_gid_filter(struct ib_device *ib_dev, u32 port, in is_ndev_for_default_gid_filter() argument
[all …]
H A Dcache.c124 static void dispatch_gid_change_event(struct ib_device *ib_dev, u32 port) in dispatch_gid_change_event() argument
128 event.device = ib_dev; in dispatch_gid_change_event()
381 static void del_gid(struct ib_device *ib_dev, u32 port, in del_gid() argument
389 dev_dbg(&ib_dev->dev, "%s port=%u index=%d gid %pI6\n", __func__, port, in del_gid()
398 if (!rdma_protocol_roce(ib_dev, port)) in del_gid()
402 if (rdma_cap_roce_gid_table(ib_dev, port)) in del_gid()
403 ib_dev->ops.del_gid(&entry->attr, &entry->context); in del_gid()
545 static int __ib_cache_gid_add(struct ib_device *ib_dev, u32 port, in __ib_cache_gid_add() argument
561 table = rdma_gid_table(ib_dev, port); in __ib_cache_gid_add()
573 attr->device = ib_dev; in __ib_cache_gid_add()
[all …]
H A Duverbs_std_types_cq.c68 struct ib_device *ib_dev = attrs->context->device; in UVERBS_HANDLER() local
81 if ((!ib_dev->ops.create_cq && !ib_dev->ops.create_cq_umem) || !ib_dev->ops.destroy_cq) in UVERBS_HANDLER()
133 !ib_dev->ops.create_cq_umem) { in UVERBS_HANDLER()
138 umem = ib_umem_get(ib_dev, buffer_va, buffer_length, IB_ACCESS_LOCAL_WRITE); in UVERBS_HANDLER()
158 !ib_dev->ops.create_cq_umem) { in UVERBS_HANDLER()
163 umem_dmabuf = ib_umem_dmabuf_get_pinned(ib_dev, buffer_offset, buffer_length, in UVERBS_HANDLER()
172 !ib_dev->ops.create_cq) { in UVERBS_HANDLER()
177 cq = rdma_zalloc_drv_obj(ib_dev, ib_cq); in UVERBS_HANDLER()
184 cq->device = ib_dev; in UVERBS_HANDLER()
194 ret = umem ? ib_dev->ops.create_cq_umem(cq, &attr, umem, attrs) : in UVERBS_HANDLER()
[all …]
H A Ddevice.c213 static void free_netdevs(struct ib_device *ib_dev);
828 pdata->ib_dev = device; in alloc_port_data()
1351 static void prevent_dealloc_device(struct ib_device *ib_dev) in prevent_dealloc_device() argument
1501 static void __ib_unregister_device(struct ib_device *ib_dev) in __ib_unregister_device() argument
1505 mutex_lock(&ib_dev->subdev_lock); in __ib_unregister_device()
1507 &ib_dev->subdev_list_head, in __ib_unregister_device()
1510 ib_dev->ops.del_sub_dev(sub); in __ib_unregister_device()
1511 ib_device_put(ib_dev); in __ib_unregister_device()
1513 mutex_unlock(&ib_dev->subdev_lock); in __ib_unregister_device()
1522 mutex_lock(&ib_dev->unregistration_lock); in __ib_unregister_device()
[all …]
H A Duverbs_std_types_dm.c56 struct ib_device *ib_dev = attrs->context->device; in UVERBS_HANDLER() local
60 if (!ib_dev->ops.alloc_dm) in UVERBS_HANDLER()
73 dm = ib_dev->ops.alloc_dm(ib_dev, attrs->context, &attr, attrs); in UVERBS_HANDLER()
77 dm->device = ib_dev; in UVERBS_HANDLER()
H A Duverbs_main.c107 if (!srcu_dereference(ufile->device->ib_dev, in ib_uverbs_get_ucontext_file()
201 struct ib_device *ib_dev; in ib_uverbs_release_file() local
207 ib_dev = srcu_dereference(file->device->ib_dev, in ib_uverbs_release_file()
209 if (ib_dev && !ib_dev->ops.disassociate_ucontext) in ib_uverbs_release_file()
210 module_put(ib_dev->ops.owner); in ib_uverbs_release_file()
494 struct ib_device *ib_dev = async_file->uobj.context->device; in ib_uverbs_init_async_event_file() local
507 INIT_IB_EVENT_HANDLER(&async_file->event_handler, ib_dev, in ib_uverbs_init_async_event_file()
933 struct ib_device *ib_dev; in ib_uverbs_open() local
945 ib_dev = srcu_dereference(dev->ib_dev, in ib_uverbs_open()
947 if (!ib_dev) { in ib_uverbs_open()
[all …]
H A Duverbs_std_types_counters.c60 struct ib_device *ib_dev = attrs->context->device; in UVERBS_HANDLER() local
69 if (!ib_dev->ops.create_counters) in UVERBS_HANDLER()
72 counters = rdma_zalloc_drv_obj(ib_dev, ib_counters); in UVERBS_HANDLER()
76 counters->device = ib_dev; in UVERBS_HANDLER()
81 ret = ib_dev->ops.create_counters(counters, attrs); in UVERBS_HANDLER()
H A Dcore_priv.h90 struct net_device *ib_device_get_netdev(struct ib_device *ib_dev,
93 void ib_enum_roce_netdev(struct ib_device *ib_dev,
129 void ib_cache_gid_set_default_gid(struct ib_device *ib_dev, u32 port,
134 int ib_cache_gid_add(struct ib_device *ib_dev, u32 port,
137 int ib_cache_gid_del(struct ib_device *ib_dev, u32 port,
140 int ib_cache_gid_del_all_netdev_gids(struct ib_device *ib_dev, u32 port,
146 unsigned long roce_gid_type_mask_support(struct ib_device *ib_dev, u32 port);
H A Duverbs_cmd.c211 struct ib_device *ib_dev; in ib_alloc_ucontext() local
213 ib_dev = srcu_dereference(ufile->device->ib_dev, in ib_alloc_ucontext()
215 if (!ib_dev) in ib_alloc_ucontext()
218 ucontext = rdma_zalloc_drv_obj(ib_dev, ib_ucontext); in ib_alloc_ucontext()
222 ucontext->device = ib_dev; in ib_alloc_ucontext()
299 struct ib_device *ib_dev; in ib_uverbs_get_context() local
311 uobj = uobj_alloc(UVERBS_OBJECT_ASYNC_EVENT, attrs, &ib_dev); in ib_uverbs_get_context()
347 struct ib_device *ib_dev = ucontext->device; in copy_query_dev_fields() local
350 resp->node_guid = ib_dev->node_guid; in copy_query_dev_fields()
386 resp->phys_port_cnt = min_t(u32, ib_dev->phys_port_cnt, U8_MAX); in copy_query_dev_fields()
[all …]
/linux/drivers/infiniband/hw/usnic/
H A Dusnic_ib_main.c79 return scnprintf(buf, buf_sz, "PF: %s ", dev_name(&vf->pf->ib_dev.dev)); in usnic_ib_dump_vf_hdr()
147 usnic_info("PF Reset on %s\n", dev_name(&us_ibdev->ib_dev.dev)); in usnic_ib_handle_usdev_event()
150 ib_event.device = &us_ibdev->ib_dev; in usnic_ib_handle_usdev_event()
158 dev_name(&us_ibdev->ib_dev.dev)); in usnic_ib_handle_usdev_event()
161 dev_name(&us_ibdev->ib_dev.dev), in usnic_ib_handle_usdev_event()
167 ib_event.device = &us_ibdev->ib_dev; in usnic_ib_handle_usdev_event()
176 dev_name(&us_ibdev->ib_dev.dev), in usnic_ib_handle_usdev_event()
182 dev_name(&us_ibdev->ib_dev.dev)); in usnic_ib_handle_usdev_event()
188 dev_name(&us_ibdev->ib_dev.dev)); in usnic_ib_handle_usdev_event()
198 container_of(ibdev, struct usnic_ib_dev, ib_dev); in usnic_ib_handle_port_event()
[all …]
H A Dusnic_ib_sysfs.c52 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); in board_id_show()
70 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); in config_show()
87 dev_name(&us_ibdev->ib_dev.dev), in config_show()
109 dev_name(&us_ibdev->ib_dev.dev)); in config_show()
122 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); in iface_show()
132 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); in max_vf_show()
142 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); in qp_per_vf_show()
156 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); in cq_per_vf_show()
255 kobject_get(&us_ibdev->ib_dev.dev.kobj); in usnic_ib_sysfs_register_usdev()
257 &us_ibdev->ib_dev.dev.kobj); in usnic_ib_sysfs_register_usdev()
[all …]
/linux/drivers/infiniband/hw/mana/
H A Dqp.c54 ibdev_dbg(&dev->ib_dev, "ind table size %u\n", 1 << log_ind_tbl_size); in mana_ib_cfg_vport_steering()
57 ibdev_dbg(&dev->ib_dev, "index %u handle 0x%llx\n", i, in mana_ib_cfg_vport_steering()
67 ibdev_dbg(&dev->ib_dev, "vport handle %llu default_rxobj 0x%llx\n", in mana_ib_cfg_vport_steering()
97 container_of(pd->device, struct mana_ib_dev, ib_dev); in mana_ib_create_qp_rss()
119 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss()
126 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss()
133 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss()
141 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss()
148 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss()
158 ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n", in mana_ib_create_qp_rss()
[all …]
H A Dmain.c15 ndev = mana_ib_get_netdev(&dev->ib_dev, port); in mana_ib_uncfg_vport()
36 ndev = mana_ib_get_netdev(&dev->ib_dev, port); in mana_ib_cfg_vport()
43 ibdev_dbg(&dev->ib_dev, in mana_ib_cfg_vport()
54 ibdev_dbg(&dev->ib_dev, "Failed to configure vPort %d\n", err); in mana_ib_cfg_vport()
63 ibdev_dbg(&dev->ib_dev, "vport handle %llx pdid %x doorbell_id %x\n", in mana_ib_cfg_vport()
80 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); in mana_ib_alloc_pd()
94 ibdev_dbg(&dev->ib_dev, in mana_ib_alloc_pd()
105 ibdev_dbg(&dev->ib_dev, "pd_handle 0x%llx pd_id %d\n", in mana_ib_alloc_pd()
123 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); in mana_ib_dealloc_pd()
134 ibdev_dbg(&dev->ib_dev, in mana_ib_dealloc_pd()
[all …]
H A Dwq.c13 container_of(pd->device, struct mana_ib_dev, ib_dev); in mana_ib_create_wq()
23 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_wq()
32 ibdev_dbg(&mdev->ib_dev, "ucmd wq_buf_addr 0x%llx\n", ucmd.wq_buf_addr); in mana_ib_create_wq()
36 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_wq()
62 struct ib_device *ib_dev = ibwq->device; in mana_ib_destroy_wq() local
65 mdev = container_of(ib_dev, struct mana_ib_dev, ib_dev); in mana_ib_destroy_wq()
H A Dmr.c66 ibdev_dbg(&dev->ib_dev, in mana_ib_gd_create_mr()
75 ibdev_dbg(&dev->ib_dev, "Failed to create mr %d, %u", err, in mana_ib_gd_create_mr()
130 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); in mana_ib_reg_user_mr()
218 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); in mana_ib_reg_user_mr_dmabuf()
284 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); in mana_ib_get_dma_mr()
314 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); in mana_ib_dereg_mr()
359 struct mana_ib_dev *dev = container_of(ibdev, struct mana_ib_dev, ib_dev); in mana_ib_alloc_dm()
401 struct mana_ib_dev *dev = container_of(ibdm->device, struct mana_ib_dev, ib_dev); in mana_ib_dealloc_dm()
417 struct mana_ib_dev *dev = container_of(ibpd->device, struct mana_ib_dev, ib_dev); in mana_ib_reg_dm_mr()
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_main.c99 net_dev = ib_device_get_netdev(&hr_dev->ib_dev, port_num); in hns_roce_get_port_state()
121 struct ib_device *ibdev = &hr_dev->ib_dev; in handle_en_event()
218 static int hns_roce_query_device(struct ib_device *ib_dev, in hns_roce_query_device() argument
222 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev); in hns_roce_query_device()
279 static int hns_roce_query_port(struct ib_device *ib_dev, u32 port_num, in hns_roce_query_port() argument
282 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev); in hns_roce_query_port()
299 ret = ib_get_eth_speed(ib_dev, port_num, &props->active_speed, in hns_roce_query_port()
302 ibdev_warn(ib_dev, "failed to get speed, ret = %d.\n", ret); in hns_roce_query_port()
304 net_dev = ib_device_get_netdev(ib_dev, port_num); in hns_roce_query_port()
306 ibdev_err(ib_dev, "find netdev %u failed!\n", port); in hns_roce_query_port()
[all …]
H A Dhns_roce_pd.c46 struct ib_device *ib_dev = ibpd->device; in hns_roce_alloc_pd() local
47 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev); in hns_roce_alloc_pd()
56 ibdev_err(ib_dev, "failed to alloc pd, id = %d.\n", id); in hns_roce_alloc_pd()
68 ibdev_err(ib_dev, "failed to copy to udata, ret = %d\n", ret); in hns_roce_alloc_pd()
93 ibdev_err(&hr_dev->ib_dev, "failed to alloc uar id(%d).\n", id); in hns_roce_uar_alloc()
128 ibdev_err(&hr_dev->ib_dev, "failed to alloc xrcdn(%d).\n", id); in hns_roce_xrcd_alloc()
H A Dhns_roce_bond.c22 return container_of(ibdev, struct hns_roce_dev, ib_dev); in hns_roce_get_hrdev_by_netdev()
103 old_dev = ib_device_get_netdev(&hr_dev->ib_dev, 1); in hns_roce_set_bond_netdev()
107 ret = ib_device_set_netdev(&hr_dev->ib_dev, active_dev, 1); in hns_roce_set_bond_netdev()
115 roce_del_all_netdev_gids(&hr_dev->ib_dev, 1, old_dev); in hns_roce_set_bond_netdev()
116 rdma_roce_rescan_port(&hr_dev->ib_dev, 1); in hns_roce_set_bond_netdev()
362 ibdev_info(&bond_grp->main_hr_dev->ib_dev, in hns_roce_set_bond()
405 ibdev_err(&bond_grp->main_hr_dev->ib_dev, in hns_roce_slave_changestate()
409 ibdev_info(&bond_grp->main_hr_dev->ib_dev, in hns_roce_slave_changestate()
447 ibdev_info(&bond_grp->main_hr_dev->ib_dev, in hns_roce_slave_change_num()
476 ib_device_put(&hr_dev->ib_dev); in hns_roce_bond_info_update_nolock()
[all …]
H A Dhns_roce_mr.c56 struct ib_device *ibdev = &hr_dev->ib_dev; in alloc_mr_key()
94 struct ib_device *ibdev = &hr_dev->ib_dev; in alloc_mr_pbl()
134 struct ib_device *ibdev = &hr_dev->ib_dev; in hns_roce_mr_free()
292 struct ib_device *ib_dev = &hr_dev->ib_dev; in hns_roce_rereg_user_mr() local
318 ibdev_warn(ib_dev, "failed to destroy MPT, ret = %d.\n", ret); in hns_roce_rereg_user_mr()
334 ibdev_err(ib_dev, "failed to alloc mr PBL, ret = %d.\n", in hns_roce_rereg_user_mr()
342 ibdev_err(ib_dev, "failed to write mtpt, ret = %d.\n", ret); in hns_roce_rereg_user_mr()
349 ibdev_err(ib_dev, "failed to create MPT, ret = %d.\n", ret); in hns_roce_rereg_user_mr()
450 struct ib_device *ibdev = &hr_dev->ib_dev; in hns_roce_map_mr_sg()
587 struct ib_device *ibdev = &hr_dev->ib_dev; in mtr_alloc_bufs()
[all …]
/linux/drivers/infiniband/ulp/isert/
H A Dib_isert.c105 struct ib_device *ib_dev = device->ib_device; in isert_create_qp() local
109 isert_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_WORKQUEUE); in isert_create_qp()
149 struct ib_device *ib_dev = device->ib_device; in isert_alloc_rx_descriptors() local
163 dma_addr = ib_dma_map_single(ib_dev, rx_desc->buf, in isert_alloc_rx_descriptors()
165 if (ib_dma_mapping_error(ib_dev, dma_addr)) in isert_alloc_rx_descriptors()
182 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr, in isert_alloc_rx_descriptors()
194 struct ib_device *ib_dev = isert_conn->device->ib_device; in isert_free_rx_descriptors() local
203 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr, in isert_free_rx_descriptors()
214 struct ib_device *ib_dev = device->ib_device; in isert_create_device_ib_res() local
218 ib_dev->attrs.max_send_sge, ib_dev->attrs.max_recv_sge); in isert_create_device_ib_res()
[all …]
/linux/drivers/infiniband/hw/mlx4/
H A Dmad.c199 ah_attr.type = rdma_ah_find_type(&dev->ib_dev, port_num); in update_sm_ah()
454 return ib_find_cached_pkey(&dev->ib_dev, port, pkey, ix); in find_slave_port_pkey_ix()
464 ret = ib_get_cached_pkey(&dev->ib_dev, port, pkey_ix, &slot_pkey); in find_slave_port_pkey_ix()
550 ret = ib_get_cached_pkey(&dev->ib_dev, port, wc->pkey_index, &cached_pkey); in mlx4_ib_send_to_slave()
577 attr.type = rdma_ah_find_type(&dev->ib_dev, port); in mlx4_ib_send_to_slave()
607 ib_dma_sync_single_for_cpu(&dev->ib_dev, in mlx4_ib_send_to_slave()
649 ib_dma_sync_single_for_device(&dev->ib_dev, in mlx4_ib_send_to_slave()
1031 ll = rdma_port_get_link_layer(&dev->ib_dev, p + 1); in mlx4_ib_mad_init()
1034 agent = ib_register_mad_agent(&dev->ib_dev, p + 1, in mlx4_ib_mad_init()
1159 mlx4_ib_warn(&dev->ib_dev, "Failed in get GUID INFO MAD_IFC\n"); in handle_slaves_guid_change()
[all …]
H A Dmain.c134 if (dev->dev.parent != ibdev->ib_dev.dev.parent || in mlx4_ib_get_netdev()
415 if (!rdma_cap_roce_gid_table(&ibdev->ib_dev, port_num)) in mlx4_ib_gid_index_to_real_index()
947 err = mlx4_ib_query_sl2vl(&mdev->ib_dev, i, &sl2vl); in mlx4_init_sl2vl_tbl()
1995 memcpy(dev->ib_dev.node_desc, out_mad->data, IB_DEVICE_NODE_DESC_MAX); in init_node_data()
2004 memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8); in init_node_data()
2016 rdma_device_to_drv_device(device, struct mlx4_ib_dev, ib_dev); in hca_type_show()
2026 rdma_device_to_drv_device(device, struct mlx4_ib_dev, ib_dev); in hw_rev_show()
2036 rdma_device_to_drv_device(device, struct mlx4_ib_dev, ib_dev); in board_id_show()
2232 ib_set_device_ops(&ibdev->ib_dev, in mlx4_ib_alloc_diag_counters()
2248 ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_hw_stats_ops); in mlx4_ib_alloc_diag_counters()
[all …]
/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_main.c103 container_of(device, struct pvrdma_dev, ib_dev); in pvrdma_get_fw_ver_str()
152 ib_event.device = &dev->ib_dev; in pvrdma_dispatch_event()
162 struct pvrdma_dev *dev = container_of(ibdev, struct pvrdma_dev, ib_dev); in pvrdma_report_event_handle()
246 dev->ib_dev.node_guid = dev->dsr->caps.node_guid; in pvrdma_register_device()
249 dev->ib_dev.num_comp_vectors = 1; in pvrdma_register_device()
250 dev->ib_dev.dev.parent = &dev->pdev->dev; in pvrdma_register_device()
252 dev->ib_dev.node_type = RDMA_NODE_IB_CA; in pvrdma_register_device()
253 dev->ib_dev.phys_port_cnt = dev->dsr->caps.phys_port_cnt; in pvrdma_register_device()
255 ib_set_device_ops(&dev->ib_dev, &pvrdma_dev_ops); in pvrdma_register_device()
272 ib_set_device_ops(&dev->ib_dev, &pvrdma_dev_srq_ops); in pvrdma_register_device()
[all …]
/linux/drivers/infiniband/sw/rxe/
H A Drxe_net.c231 ib_device_put(&rxe->ib_dev); in rxe_udp_encap_recv()
480 if (WARN_ON(!ib_device_try_get(&pkt->rxe->ib_dev))) { in rxe_loopback()
540 attr = rdma_get_gid_attr(&rxe->ib_dev, port_num, av->grh.sgid_index); in rxe_init_packet()
598 ndev = rxe_ib_device_get_netdev(&rxe->ib_dev); in rxe_parent_name()
612 rxe = ib_alloc_device(rxe_dev, ib_dev); in rxe_net_add()
616 ib_mark_name_assigned_by_user(&rxe->ib_dev); in rxe_net_add()
620 ib_dealloc_device(&rxe->ib_dev); in rxe_net_add()
632 ev.device = &rxe->ib_dev; in rxe_port_event()
643 dev_info(&rxe->ib_dev.dev, "set active\n"); in rxe_port_up()
651 dev_info(&rxe->ib_dev.dev, "set down\n"); in rxe_port_down()
[all …]

1234