Searched refs:drm_priv (Results 1 – 8 of 8) sorted by relevance
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_amdkfd_gpuvm.c | 1621 uint64_t amdgpu_amdkfd_gpuvm_get_process_page_dir(void *drm_priv) in amdgpu_amdkfd_gpuvm_get_process_page_dir() argument 1623 struct amdgpu_vm *avm = drm_priv_to_vm(drm_priv); in amdgpu_amdkfd_gpuvm_get_process_page_dir() 1711 void *drm_priv, struct kgd_mem **mem, in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() argument 1714 struct amdgpu_vm *avm = drm_priv_to_vm(drm_priv); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1820 ret = drm_vma_node_allow(&gobj->vma_node, drm_priv); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1881 drm_vma_node_revoke(&gobj->vma_node, drm_priv); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1903 struct amdgpu_device *adev, struct kgd_mem *mem, void *drm_priv, in amdgpu_amdkfd_gpuvm_free_memory_of_gpu() argument 1991 drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv); in amdgpu_amdkfd_gpuvm_free_memory_of_gpu() 2018 void *drm_priv) in amdgpu_amdkfd_gpuvm_map_memory_to_gpu() argument 2020 struct amdgpu_vm *avm = drm_priv_to_vm(drm_priv); in amdgpu_amdkfd_gpuvm_map_memory_to_gpu() [all …]
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_chardev.c | 1166 pdd->drm_priv, (struct kgd_mem **) &mem, &offset, in kfd_ioctl_alloc_memory_of_gpu() 1203 pdd->drm_priv, NULL); in kfd_ioctl_alloc_memory_of_gpu() 1246 (struct kgd_mem *)mem, pdd->drm_priv, &size); in kfd_ioctl_free_memory_of_gpu() 1334 peer_pdd->drm_priv); in kfd_ioctl_map_memory_to_gpu() 1434 peer_pdd->dev->adev, (struct kgd_mem *)mem, peer_pdd->drm_priv); in kfd_ioctl_unmap_memory_from_gpu() 1461 err = amdgpu_amdkfd_gpuvm_dmaunmap_mem(mem, peer_pdd->drm_priv); in kfd_ioctl_unmap_memory_from_gpu() 1604 args->va_addr, pdd->drm_priv, in kfd_ioctl_import_dmabuf() 1624 pdd->drm_priv, NULL); in kfd_ioctl_import_dmabuf() 2020 if (amdgpu_amdkfd_bo_mapped_to_dev(p->pdds[i]->drm_priv, kgd_mem)) in criu_checkpoint_bos() 2367 bo_bucket->size, pdd->drm_priv, kgd_mem, in criu_restore_memory_of_gpu() [all …]
|
| H A D | kfd_smi_events.c | 352 if (!pdd->drm_priv) in kfd_smi_event_process() 355 avm = drm_priv_to_vm(pdd->drm_priv); in kfd_smi_event_process()
|
| H A D | kfd_queue.c | 257 vm = drm_priv_to_vm(pdd->drm_priv); in kfd_queue_acquire_buffers() 383 vm = drm_priv_to_vm(pdd->drm_priv); in kfd_queue_unref_bo_vas()
|
| H A D | kfd_svm.c | 1415 drm_priv_to_vm(pdd->drm_priv), in svm_range_unmap_from_gpus() 1440 struct amdgpu_vm *vm = drm_priv_to_vm(pdd->drm_priv); in svm_range_map_to_gpu() 1604 vm = drm_priv_to_vm(pdd->drm_priv); in svm_range_reserve_bos() 1624 drm_priv_to_vm(pdd->drm_priv), NULL, in svm_range_reserve_bos() 2880 if (!p->pdds[i]->drm_priv) in svm_range_check_vm_userptr() 2883 vm = drm_priv_to_vm(p->pdds[i]->drm_priv); in svm_range_check_vm_userptr() 3427 if (!p->pdds[i]->drm_priv) in svm_range_check_vm() 3430 vm = drm_priv_to_vm(p->pdds[i]->drm_priv); in svm_range_check_vm()
|
| H A D | kfd_priv.h | 775 void *drm_priv; member 1560 struct amdgpu_vm *vm = drm_priv_to_vm(pdd->drm_priv); in kfd_flush_tlb()
|
| H A D | kfd_device_queue_manager.c | 1208 if (!pdd->drm_priv) in evict_process_queues_cpsch() 1265 pd_base = amdgpu_amdkfd_gpuvm_get_process_page_dir(pdd->drm_priv); in restore_process_queues_nocpsch() 1356 if (!pdd->drm_priv) in restore_process_queues_cpsch() 1363 qpd->page_table_base = amdgpu_amdkfd_gpuvm_get_process_page_dir(pdd->drm_priv); in restore_process_queues_cpsch() 1412 pd_base = amdgpu_amdkfd_gpuvm_get_process_page_dir(pdd->drm_priv); in register_process()
|
| H A D | kfd_process_queue_manager.c | 587 vm = drm_priv_to_vm(pdd->drm_priv); in pqm_update_queue_properties()
|