Lines Matching refs:gmc
649 if (!adev->gmc.noretry && !amdgpu_passthrough(adev)) in amdgpu_device_detect_runtime_pm_mode()
763 last = min(pos + size, adev->gmc.visible_vram_size); in amdgpu_device_aper_access()
1117 int rbar_size = pci_rebar_bytes_to_size(adev->gmc.real_vram_size); in amdgpu_device_resize_fb_bar()
1148 if (adev->gmc.real_vram_size && in amdgpu_device_resize_fb_bar()
1149 (pci_resource_len(adev->pdev, 0) >= adev->gmc.real_vram_size)) in amdgpu_device_resize_fb_bar()
2102 if (adev->gmc.xgmi.supported) in amdgpu_device_ip_early_init()
2434 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_init()
2709 ((adev->asic_type == CHIP_ARCTURUS && adev->gmc.xgmi.num_physical_nodes > 1) || in amdgpu_device_ip_late_init()
2713 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_late_init()
2729 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_ip_late_init()
2846 if ((adev->flags & AMD_IS_APU) && !adev->gmc.is_app_apu && in amdgpu_device_ip_fini_early()
2876 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini()
3699 adev->gmc.gart_size = 512 * 1024 * 1024; in amdgpu_device_init()
3707 adev->gmc.gmc_funcs = NULL; in amdgpu_device_init()
3890 if (adev->gmc.xgmi.supported) { in amdgpu_device_init()
3899 if (adev->gmc.xgmi.connected_to_cpu) { in amdgpu_device_init()
3919 (adev->gmc.xgmi.connected_to_cpu && in amdgpu_device_init()
3951 if (adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_init()
4166 if (!adev->gmc.xgmi.connected_to_cpu && !adev->gmc.is_app_apu) { in amdgpu_device_unmap_mmio()
4167 arch_phys_wc_del(adev->gmc.vram_mtrr); in amdgpu_device_unmap_mmio()
4168 arch_io_free_memtype_wc(adev->gmc.aper_base, adev->gmc.aper_size); in amdgpu_device_unmap_mmio()
4550 unsigned int prev_physical_node_id = adev->gmc.xgmi.physical_node_id; in amdgpu_virt_resume()
4564 prev_physical_node_id, adev->gmc.xgmi.physical_node_id); in amdgpu_virt_resume()
4568 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in amdgpu_virt_resume()
4891 if (hive && adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_reset_sriov()
5248 tmp_adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_reinit_after_reset()
5281 tmp_adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_reinit_after_reset()
5340 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
5359 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
5517 if (!amdgpu_sriov_vf(adev) && (adev->gmc.xgmi.num_physical_nodes > 1) && hive) { in amdgpu_device_recovery_prepare()
5518 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_device_recovery_prepare()
6186 !adev->gmc.xgmi.connected_to_cpu && in amdgpu_device_is_peer_accessible()
6192 bool is_large_bar = adev->gmc.visible_vram_size && in amdgpu_device_is_peer_accessible()
6193 adev->gmc.real_vram_size == adev->gmc.visible_vram_size; in amdgpu_device_is_peer_accessible()
6200 adev->gmc.aper_base + adev->gmc.aper_size - 1; in amdgpu_device_is_peer_accessible()
6202 p2p_addressable = !(adev->gmc.aper_base & address_mask || in amdgpu_device_is_peer_accessible()
6365 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_pci_slot_reset()
6413 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_pci_slot_reset()
6469 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_pci_resume()
6603 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_device_flush_hdp()
6626 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_device_invalidate_hdp()