Home
last modified time | relevance | path

Searched refs:peer_dev (Results 1 – 9 of 9) sorted by relevance

/linux/drivers/infiniband/hw/mlx5/
H A Dib_rep.c33 struct mlx5_core_dev *peer_dev; in mlx5_ib_num_ports_update() local
36 mlx5_lag_for_each_peer_mdev(dev, peer_dev, i) { in mlx5_ib_num_ports_update()
37 u32 peer_num_ports = mlx5_eswitch_get_total_vports(peer_dev); in mlx5_ib_num_ports_update()
39 if (mlx5_lag_is_mpesw(peer_dev)) in mlx5_ib_num_ports_update()
78 struct mlx5_core_dev *peer_dev; in mlx5_ib_release_transport() local
81 mlx5_lag_for_each_peer_mdev(dev, peer_dev, i) { in mlx5_ib_release_transport()
82 ret = mlx5_ib_set_owner_transport(peer_dev, peer_dev); in mlx5_ib_release_transport()
89 struct mlx5_core_dev *peer_dev; in mlx5_ib_take_transport() local
93 mlx5_lag_for_each_peer_mdev(dev, peer_dev, i) { in mlx5_ib_take_transport()
94 ret = mlx5_ib_set_owner_transport(peer_dev, dev); in mlx5_ib_take_transport()
[all …]
/linux/drivers/net/netdevsim/
H A Dnetdev.c43 struct net_device *peer_dev; in nsim_start_peer_tx_queue() local
55 peer_dev = peer_ns->netdev; in nsim_start_peer_tx_queue()
56 if (dev->real_num_tx_queues != peer_dev->num_rx_queues) in nsim_start_peer_tx_queue()
59 txq = netdev_get_tx_queue(peer_dev, idx); in nsim_start_peer_tx_queue()
124 struct net_device *peer_dev; in nsim_start_xmit() local
139 peer_dev = dev; in nsim_start_xmit()
144 peer_dev = peer_ns->netdev; in nsim_start_xmit()
152 if (rxq >= peer_dev->num_rx_queues) in nsim_start_xmit()
153 rxq = rxq % peer_dev->num_rx_queues; in nsim_start_xmit()
156 cfg = peer_dev->cfg; in nsim_start_xmit()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/lib/
H A Dclock.c1437 struct mlx5_core_dev *peer_dev, *next = NULL; in mlx5_shared_clock_register() local
1454 mlx5_devcom_for_each_peer_entry(mdev->clock_state->compdev, peer_dev, pos) { in mlx5_shared_clock_register()
1455 if (peer_dev->clock) { in mlx5_shared_clock_register()
1456 next = peer_dev; in mlx5_shared_clock_register()
1478 struct mlx5_core_dev *peer_dev, *next = NULL; in mlx5_shared_clock_unregister() local
1483 mlx5_devcom_for_each_peer_entry(mdev->clock_state->compdev, peer_dev, pos) { in mlx5_shared_clock_unregister()
1484 if (peer_dev->clock && peer_dev != mdev) { in mlx5_shared_clock_unregister()
1485 next = peer_dev; in mlx5_shared_clock_unregister()
1561 struct mlx5_core_dev *peer_dev, *next = NULL; in mlx5_clock_unload() local
1574 mlx5_devcom_for_each_peer_entry(mdev->clock_state->compdev, peer_dev, pos) { in mlx5_clock_unload()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Deswitch_offloads.c1128 struct mlx5_core_dev *peer_dev, in peer_miss_rules_setup() argument
1146 MLX5_CAP_GEN(peer_dev, vhca_id)); in peer_miss_rules_setup()
1158 dest->vport.num = peer_dev->priv.eswitch->manager_vport; in peer_miss_rules_setup()
1159 dest->vport.vhca_id = MLX5_CAP_GEN(peer_dev, vhca_id); in peer_miss_rules_setup()
1184 struct mlx5_core_dev *peer_dev) in esw_add_fdb_peer_miss_rules() argument
1186 struct mlx5_eswitch *peer_esw = peer_dev->priv.eswitch; in esw_add_fdb_peer_miss_rules()
1197 if (!MLX5_VPORT_MANAGER(peer_dev) && in esw_add_fdb_peer_miss_rules()
1198 !mlx5_core_is_ecpf_esw_manager(peer_dev)) in esw_add_fdb_peer_miss_rules()
1205 peer_miss_rules_setup(esw, peer_dev, spec, &dest); in esw_add_fdb_peer_miss_rules()
1217 if (mlx5_core_is_ecpf_esw_manager(peer_dev) && in esw_add_fdb_peer_miss_rules()
[all …]
H A Dfs_cmd.c252 struct mlx5_core_dev *peer_dev; in mlx5_cmd_update_root_ft() local
255 mlx5_lag_for_each_peer_mdev(dev, peer_dev, i) { in mlx5_cmd_update_root_ft()
256 err = mlx5_cmd_set_slave_root_fdb(dev, peer_dev, !disconnect, in mlx5_cmd_update_root_ft()
259 mlx5_lag_for_each_peer_mdev(dev, peer_dev, j) { in mlx5_cmd_update_root_ft()
261 mlx5_cmd_set_slave_root_fdb(dev, peer_dev, 1, in mlx5_cmd_update_root_ft()
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_topology.c1318 struct kfd_topology_device *peer_dev; in kfd_fill_iolink_non_crat_info() local
1327 peer_dev = kfd_topology_device_by_proximity_domain( in kfd_fill_iolink_non_crat_info()
1330 if (!peer_dev) in kfd_fill_iolink_non_crat_info()
1334 if (!peer_dev->gpu && in kfd_fill_iolink_non_crat_info()
1342 peer_dev->node_props.hive_id = dev->node_props.hive_id; in kfd_fill_iolink_non_crat_info()
1345 list_for_each_entry(inbound_link, &peer_dev->io_link_props, in kfd_fill_iolink_non_crat_info()
1351 kfd_set_iolink_no_atomics(peer_dev, dev, inbound_link); in kfd_fill_iolink_non_crat_info()
1352 kfd_set_iolink_non_coherent(peer_dev, link, inbound_link); in kfd_fill_iolink_non_crat_info()
1353 kfd_set_recommended_sdma_engines(peer_dev, link, inbound_link); in kfd_fill_iolink_non_crat_info()
1361 peer_dev = kfd_topology_device_by_proximity_domain( in kfd_fill_iolink_non_crat_info()
[all …]
H A Dkfd_crat.c2243 struct kfd_topology_device *peer_dev; in kfd_create_vcrat_image_gpu() local
2361 peer_dev = kfd_topology_device_by_proximity_domain_no_lock(nid); in kfd_create_vcrat_image_gpu()
2362 if (!peer_dev || !peer_dev->gpu) in kfd_create_vcrat_image_gpu()
2364 if (peer_dev->gpu->kfd->hive_id != kdev->kfd->hive_id) in kfd_create_vcrat_image_gpu()
2366 if (!amdgpu_xgmi_get_is_sharing_enabled(kdev->adev, peer_dev->gpu->adev)) in kfd_create_vcrat_image_gpu()
2372 &avail_size, kdev, peer_dev->gpu, in kfd_create_vcrat_image_gpu()
/linux/drivers/net/
H A Dveth.c967 struct net_device *peer_dev; in veth_poll() local
974 peer_dev = rcu_dereference_check(priv->peer, rcu_read_lock_bh_held()); in veth_poll()
975 peer_txq = peer_dev ? netdev_get_tx_queue(peer_dev, queue_idx) : NULL; in veth_poll()
/linux/drivers/net/ethernet/mellanox/mlx5/core/lag/
H A Dlag.c1939 struct mlx5_core_dev *peer_dev = NULL; in mlx5_lag_get_next_peer_mdev() local
1961 peer_dev = ldev->pf[idx].dev; in mlx5_lag_get_next_peer_mdev()
1965 return peer_dev; in mlx5_lag_get_next_peer_mdev()