Lines Matching defs:tmp_adev
954 struct amdgpu_device *tmp_adev;
957 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
958 ret = psp_xgmi_initialize(&tmp_adev->psp, set_extended_data, false);
960 dev_err(tmp_adev->dev,
976 struct amdgpu_device *tmp_adev = NULL;
1030 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
1032 if (tmp_adev != adev) {
1033 top_info = &tmp_adev->psp.xgmi_context.top_info;
1038 ret = amdgpu_xgmi_update_topology(hive, tmp_adev);
1058 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
1059 ret = psp_xgmi_get_topology_info(&tmp_adev->psp, count,
1060 &tmp_adev->psp.xgmi_context.top_info, false);
1062 dev_err(tmp_adev->dev,
1064 tmp_adev->gmc.xgmi.node_id,
1065 tmp_adev->gmc.xgmi.hive_id, ret);
1081 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
1082 ret = psp_xgmi_get_topology_info(&tmp_adev->psp, count,
1083 &tmp_adev->psp.xgmi_context.top_info, true);
1085 dev_err(tmp_adev->dev,
1087 tmp_adev->gmc.xgmi.node_id,
1088 tmp_adev->gmc.xgmi.hive_id, ret);
1631 struct amdgpu_device *tmp_adev;
1638 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head)
1639 list_add_tail(&tmp_adev->reset_list, &device_list);
1641 tmp_adev = list_first_entry(&device_list, struct amdgpu_device,
1643 amdgpu_device_lock_reset_domain(tmp_adev->reset_domain);
1646 reset_context.reset_req_dev = tmp_adev;
1654 amdgpu_device_unlock_reset_domain(tmp_adev->reset_domain);
1656 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
1657 r = amdgpu_ras_init_badpage_info(tmp_adev);
1659 dev_err(tmp_adev->dev,
1702 struct amdgpu_device *tmp_adev;
1718 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
1720 tmp_adev, req_nps_mode);
1727 adev->gmc.gmc_funcs->query_mem_partition_mode(tmp_adev);
1729 tmp_adev, &hive->device_list, gmc.xgmi.head)
1731 tmp_adev, cur_nps_mode);