Lines Matching refs:gmc

299 	if (adev->gmc.xgmi.num_physical_nodes <= 1)  in amdgpu_xgmi_get_ext_link()
350 if (adev->gmc.xgmi.num_physical_nodes <= 1) in amdgpu_get_xgmi_link_status()
450 return sysfs_emit(buf, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id()
461 return sysfs_emit(buf, "%u\n", adev->gmc.xgmi.physical_node_id); in amdgpu_xgmi_show_physical_id()
510 if (top->nodes[i].node_id == adev->gmc.xgmi.node_id) { in amdgpu_xgmi_show_connected_port_num()
678 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
689 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
741 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
813 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
814 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
845 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
846 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
865 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_get_hops_count()
869 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count()
873 peer_adev->gmc.xgmi.physical_node_id); in amdgpu_xgmi_get_hops_count()
884 int num_lanes = adev->gmc.xgmi.max_width; in amdgpu_xgmi_get_bandwidth()
885 int speed = adev->gmc.xgmi.max_speed; in amdgpu_xgmi_get_bandwidth()
894 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_get_bandwidth()
905 if (top->nodes[i].node_id != peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_bandwidth()
915 peer_adev->gmc.xgmi.physical_node_id); in amdgpu_xgmi_get_bandwidth()
937 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_is_sharing_enabled()
955 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_initialize_hive_get_data_partition()
978 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_add_device()
989 ret = psp_xgmi_get_hive_id(&adev->psp, &adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
996 ret = psp_xgmi_get_node_id(&adev->psp, &adev->gmc.xgmi.node_id); in amdgpu_xgmi_add_device()
1003 adev->gmc.xgmi.hive_id = 16; in amdgpu_xgmi_add_device()
1004 adev->gmc.xgmi.node_id = adev->gmc.xgmi.physical_node_id + 16; in amdgpu_xgmi_add_device()
1012 adev->gmc.xgmi.node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
1019 list_add_tail(&adev->gmc.xgmi.head, &hive->device_list); in amdgpu_xgmi_add_device()
1028 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
1033 adev->gmc.xgmi.node_id; in amdgpu_xgmi_add_device()
1049 adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
1050 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
1056 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
1062 tmp_adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
1063 tmp_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
1079 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
1085 tmp_adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
1086 tmp_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
1108 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
1112 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id, in amdgpu_xgmi_add_device()
1123 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_remove_device()
1134 list_del(&adev->gmc.xgmi.head); in amdgpu_xgmi_remove_device()
1209 if (!adev->gmc.xgmi.supported || in amdgpu_xgmi_ras_late_init()
1210 adev->gmc.xgmi.num_physical_nodes == 0) in amdgpu_xgmi_ras_late_init()
1242 struct amdgpu_xgmi *xgmi = &adev->gmc.xgmi; in amdgpu_xgmi_get_relative_phy_addr()
1606 if (!adev->gmc.xgmi.ras) in amdgpu_xgmi_ras_sw_init()
1609 ras = adev->gmc.xgmi.ras; in amdgpu_xgmi_ras_sw_init()
1619 adev->gmc.xgmi.ras_if = &ras->ras_block.ras_comm; in amdgpu_xgmi_ras_sw_init()
1636 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_xgmi_reset_on_init_work()
1654 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_reset_on_init_work()
1682 if (num_devs == adev->gmc.xgmi.num_physical_nodes) { in amdgpu_xgmi_reset_on_init()
1716 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_request_nps_change()
1717 r = adev->gmc.gmc_funcs->request_mem_partition_mode( in amdgpu_xgmi_request_nps_change()
1725 adev->gmc.gmc_funcs->query_mem_partition_mode(tmp_adev); in amdgpu_xgmi_request_nps_change()
1727 tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_xgmi_request_nps_change()
1728 adev->gmc.gmc_funcs->request_mem_partition_mode( in amdgpu_xgmi_request_nps_change()
1742 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive()
1743 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
1748 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_early_init()
1756 adev->gmc.xgmi.max_speed = 25; in amdgpu_xgmi_early_init()
1757 adev->gmc.xgmi.max_width = 16; in amdgpu_xgmi_early_init()
1763 adev->gmc.xgmi.max_speed = 32; in amdgpu_xgmi_early_init()
1764 adev->gmc.xgmi.max_width = 16; in amdgpu_xgmi_early_init()
1774 adev->gmc.xgmi.max_speed = max_speed; in amgpu_xgmi_set_max_speed_width()
1775 adev->gmc.xgmi.max_width = max_width; in amgpu_xgmi_set_max_speed_width()