Searched refs:peer_dev (Results 1 – 7 of 7) sorted by relevance
/linux/drivers/infiniband/hw/mlx5/ |
H A D | ib_rep.c | 33 struct mlx5_core_dev *peer_dev; in mlx5_ib_num_ports_update() local 36 mlx5_lag_for_each_peer_mdev(dev, peer_dev, i) { in mlx5_ib_num_ports_update() 37 u32 peer_num_ports = mlx5_eswitch_get_total_vports(peer_dev); in mlx5_ib_num_ports_update() 39 if (mlx5_lag_is_mpesw(peer_dev)) in mlx5_ib_num_ports_update() 53 struct mlx5_core_dev *peer_dev; in mlx5_ib_vport_rep_load() local 71 mlx5_lag_for_each_peer_mdev(dev, peer_dev, i) { in mlx5_ib_vport_rep_load() 72 u32 peer_n_ports = mlx5_eswitch_get_total_vports(peer_dev); in mlx5_ib_vport_rep_load() 74 if (mlx5_lag_is_master(peer_dev)) in mlx5_ib_vport_rep_load() 75 lag_master = peer_dev; in mlx5_ib_vport_rep_load() 80 if (mlx5_get_dev_index(peer_dev) < mlx5_get_dev_index(dev)) in mlx5_ib_vport_rep_load()
|
/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_topology.c | 1321 struct kfd_topology_device *peer_dev; in kfd_fill_iolink_non_crat_info() local 1330 peer_dev = kfd_topology_device_by_proximity_domain( in kfd_fill_iolink_non_crat_info() 1333 if (!peer_dev) in kfd_fill_iolink_non_crat_info() 1337 if (!peer_dev->gpu && in kfd_fill_iolink_non_crat_info() 1345 peer_dev->node_props.hive_id = dev->node_props.hive_id; in kfd_fill_iolink_non_crat_info() 1348 list_for_each_entry(inbound_link, &peer_dev->io_link_props, in kfd_fill_iolink_non_crat_info() 1354 kfd_set_iolink_no_atomics(peer_dev, dev, inbound_link); in kfd_fill_iolink_non_crat_info() 1355 kfd_set_iolink_non_coherent(peer_dev, link, inbound_link); in kfd_fill_iolink_non_crat_info() 1356 kfd_set_recommended_sdma_engines(peer_dev, link, inbound_link); in kfd_fill_iolink_non_crat_info() 1364 peer_dev = kfd_topology_device_by_proximity_domain( in kfd_fill_iolink_non_crat_info() [all …]
|
H A D | kfd_crat.c | 2209 struct kfd_topology_device *peer_dev; in kfd_create_vcrat_image_gpu() local 2327 peer_dev = kfd_topology_device_by_proximity_domain_no_lock(nid); in kfd_create_vcrat_image_gpu() 2328 if (!peer_dev->gpu) in kfd_create_vcrat_image_gpu() 2330 if (peer_dev->gpu->kfd->hive_id != kdev->kfd->hive_id) in kfd_create_vcrat_image_gpu() 2336 &avail_size, kdev, peer_dev->gpu, in kfd_create_vcrat_image_gpu()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | fs_cmd.c | 247 struct mlx5_core_dev *peer_dev; in mlx5_cmd_update_root_ft() local 250 mlx5_lag_for_each_peer_mdev(dev, peer_dev, i) { in mlx5_cmd_update_root_ft() 251 err = mlx5_cmd_set_slave_root_fdb(dev, peer_dev, !disconnect, in mlx5_cmd_update_root_ft() 254 mlx5_lag_for_each_peer_mdev(dev, peer_dev, j) { in mlx5_cmd_update_root_ft() 256 mlx5_cmd_set_slave_root_fdb(dev, peer_dev, 1, in mlx5_cmd_update_root_ft()
|
H A D | eswitch_offloads.c | 1129 struct mlx5_core_dev *peer_dev, in peer_miss_rules_setup() argument 1147 MLX5_CAP_GEN(peer_dev, vhca_id)); in peer_miss_rules_setup() 1159 dest->vport.num = peer_dev->priv.eswitch->manager_vport; in peer_miss_rules_setup() 1160 dest->vport.vhca_id = MLX5_CAP_GEN(peer_dev, vhca_id); in peer_miss_rules_setup() 1185 struct mlx5_core_dev *peer_dev) in esw_add_fdb_peer_miss_rules() argument 1206 peer_miss_rules_setup(esw, peer_dev, spec, &dest); in esw_add_fdb_peer_miss_rules() 1220 esw_set_peer_miss_rule_source_port(esw, peer_dev->priv.eswitch, in esw_add_fdb_peer_miss_rules() 1246 peer_dev->priv.eswitch, in esw_add_fdb_peer_miss_rules() 1260 if (i >= mlx5_core_max_ec_vfs(peer_dev)) in esw_add_fdb_peer_miss_rules() 1262 esw_set_peer_miss_rule_source_port(esw, peer_dev->priv.eswitch, in esw_add_fdb_peer_miss_rules() [all …]
|
H A D | mlx5_core.h | 357 bool mlx5_same_hw_devs(struct mlx5_core_dev *dev, struct mlx5_core_dev *peer_dev);
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/lag/ |
H A D | lag.c | 1575 struct mlx5_core_dev *peer_dev = NULL; in mlx5_lag_get_next_peer_mdev() local 1597 peer_dev = ldev->pf[idx].dev; in mlx5_lag_get_next_peer_mdev() 1601 return peer_dev; in mlx5_lag_get_next_peer_mdev()
|