| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_vm.c | 294 struct radeon_bo_va *bo_va; in radeon_vm_bo_find() local 296 list_for_each_entry(bo_va, &bo->va, bo_list) { in radeon_vm_bo_find() 297 if (bo_va->vm == vm) in radeon_vm_bo_find() 298 return bo_va; in radeon_vm_bo_find() 321 struct radeon_bo_va *bo_va; in radeon_vm_bo_add() local 323 bo_va = kzalloc_obj(struct radeon_bo_va); in radeon_vm_bo_add() 324 if (bo_va == NULL) in radeon_vm_bo_add() 327 bo_va->vm = vm; in radeon_vm_bo_add() 328 bo_va->bo = bo; in radeon_vm_bo_add() 329 bo_va->it.start = 0; in radeon_vm_bo_add() [all …]
|
| H A D | radeon_trace.h | 66 TP_PROTO(struct radeon_bo_va *bo_va), 67 TP_ARGS(bo_va), 75 __entry->soffset = bo_va->it.start; 76 __entry->eoffset = bo_va->it.last + 1; 77 __entry->flags = bo_va->flags;
|
| H A D | radeon_cs.c | 505 struct radeon_bo_va *bo_va; in radeon_bo_vm_update_pte() local 530 bo_va = radeon_vm_bo_find(vm, bo); in radeon_bo_vm_update_pte() 531 if (bo_va == NULL) { in radeon_bo_vm_update_pte() 536 r = radeon_vm_bo_update(rdev, bo_va, bo->tbo.resource); in radeon_bo_vm_update_pte() 540 radeon_sync_fence(&p->ib.sync, bo_va->last_pt_update); in radeon_bo_vm_update_pte()
|
| /linux/drivers/gpu/drm/lima/ |
| H A D | lima_vm.c | 79 struct lima_bo_va *bo_va, *ret = NULL; in lima_vm_bo_find() local 81 list_for_each_entry(bo_va, &bo->va, list) { in lima_vm_bo_find() 82 if (bo_va->vm == vm) { in lima_vm_bo_find() 83 ret = bo_va; in lima_vm_bo_find() 93 struct lima_bo_va *bo_va; in lima_vm_bo_add() local 99 bo_va = lima_vm_bo_find(vm, bo); in lima_vm_bo_add() 100 if (bo_va) { in lima_vm_bo_add() 101 bo_va->ref_count++; in lima_vm_bo_add() 112 bo_va = kzalloc_obj(*bo_va); in lima_vm_bo_add() 113 if (!bo_va) { in lima_vm_bo_add() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_vm.c | 493 struct amdgpu_bo_va *bo_va; in amdgpu_vm_lock_done_list() local 500 bo_va = list_entry(prev->next, typeof(*bo_va), base.vm_status); in amdgpu_vm_lock_done_list() 502 bo = bo_va->base.bo; in amdgpu_vm_lock_done_list() 1280 int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va, in amdgpu_vm_bo_update() argument 1283 struct amdgpu_bo *bo = bo_va->base.bo; in amdgpu_vm_bo_update() 1284 struct amdgpu_vm *vm = bo_va->base.vm; in amdgpu_vm_bo_update() 1320 if (drm_gem_is_imported(obj) && bo_va->is_xgmi) { in amdgpu_vm_bo_update() 1361 last_update = &bo_va->last_pt_update; in amdgpu_vm_bo_update() 1363 if (!clear && bo_va->base.moved) { in amdgpu_vm_bo_update() 1365 list_splice_init(&bo_va->valids, &bo_va->invalids); in amdgpu_vm_bo_update() [all …]
|
| H A D | amdgpu_amdkfd_gpuvm.c | 83 if (entry->bo_va->base.vm == avm) in kfd_mem_is_attached() 567 struct amdgpu_bo *bo = attachment->bo_va->base.bo; in kfd_mem_dmamap_userptr() 614 struct amdgpu_bo *bo = attachment->bo_va->base.bo; in kfd_mem_dmamap_dmabuf() 651 struct amdgpu_bo *bo = attachment->bo_va->base.bo; in kfd_mem_dmamap_sg_bo() 728 struct amdgpu_bo *bo = attachment->bo_va->base.bo; in kfd_mem_dmaunmap_userptr() 774 struct amdgpu_bo *bo = attachment->bo_va->base.bo; in kfd_mem_dmaunmap_sg_bo() 879 struct amdgpu_bo_va *bo_va; in kfd_mem_attach() local 974 bo_va = amdgpu_vm_bo_find(vm, bo[i]); in kfd_mem_attach() 975 if (!bo_va) in kfd_mem_attach() 976 bo_va = amdgpu_vm_bo_add(adev, vm, bo[i]); in kfd_mem_attach() [all …]
|
| H A D | amdgpu_vm.h | 541 struct amdgpu_bo_va *bo_va, 557 struct amdgpu_bo_va *bo_va, 561 struct amdgpu_bo_va *bo_va, 565 struct amdgpu_bo_va *bo_va, 574 struct amdgpu_bo_va *bo_va); 694 #define amdgpu_vm_bo_va_for_each_valid_mapping(bo_va, mapping) \ argument 695 list_for_each_entry(mapping, &(bo_va)->valids, list) 696 #define amdgpu_vm_bo_va_for_each_invalid_mapping(bo_va, mapping) \ argument 697 list_for_each_entry(mapping, &(bo_va)->invalids, list)
|
| H A D | amdgpu_csa.h | 35 struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va, 38 struct amdgpu_bo *bo, struct amdgpu_bo_va *bo_va,
|
| H A D | amdgpu_userq.c | 239 atomic_set(&va_map->bo_va->userq_va_mapped, 1); in amdgpu_userq_buffer_va_list_add() 286 if (!IS_ERR_OR_NULL(mapping) && atomic_read(&mapping->bo_va->userq_va_mapped)) in amdgpu_userq_buffer_va_mapped() 315 atomic_set(&mapping->bo_va->userq_va_mapped, 0); in amdgpu_userq_buffer_va_list_del() 1058 struct amdgpu_bo_va *bo_va; in amdgpu_userq_bo_validate() local 1064 bo_va = list_first_entry(&vm->invalidated, in amdgpu_userq_bo_validate() 1069 bo = bo_va->base.bo; in amdgpu_userq_bo_validate() 1080 ret = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_userq_bo_validate() 1102 struct amdgpu_bo_va *bo_va; in amdgpu_userq_vm_validate() local 1161 list_for_each_entry(bo_va, &vm->done, base.vm_status) { in amdgpu_userq_vm_validate() 1162 bo = bo_va->base.bo; in amdgpu_userq_vm_validate() [all …]
|
| H A D | mes_userqueue.c | 61 obj = wptr_mapping->bo_va->base.bo; in mes_userq_create_wptr_mapping() 68 wptr_obj->obj = amdgpu_bo_ref(wptr_mapping->bo_va->base.bo); in mes_userq_create_wptr_mapping()
|
| H A D | mes_v12_1.c | 1991 struct amdgpu_bo_va **bo_va, u64 va, int size) in mes_v12_1_map_test_bo() argument 1996 r = amdgpu_map_static_csa(adev, vm, bo, bo_va, va, size); in mes_v12_1_map_test_bo() 2002 r = amdgpu_vm_bo_update(adev, *bo_va, false); in mes_v12_1_map_test_bo() 2007 amdgpu_sync_fence(&sync, (*bo_va)->last_pt_update, GFP_KERNEL); in mes_v12_1_map_test_bo() 2225 struct amdgpu_bo_va *bo_va = NULL; in mes_v12_1_self_test() local 2257 r = mes_v12_1_map_test_bo(adev, ctx_bo, vm, &bo_va, in mes_v12_1_self_test() 2271 amdgpu_unmap_static_csa(adev, vm, ctx_bo, bo_va, USER_CTX_VA); in mes_v12_1_self_test()
|
| H A D | amdgpu_dev_coredump.c | 373 abo = amdgpu_bo_ref(mapping->bo_va->base.bo); in amdgpu_devcoredump_format()
|
| H A D | amdgpu_userq_fence.c | 376 bo = amdgpu_bo_ref(mapping->bo_va->base.bo); in amdgpu_userq_fence_read_wptr()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_queue.c | 218 *pbo = amdgpu_bo_ref(mapping->bo_va->base.bo); in kfd_queue_buffer_get() 219 mapping->bo_va->queue_refcount++; in kfd_queue_buffer_get() 369 struct amdgpu_bo_va *bo_va; in kfd_queue_unref_bo_va() local 371 bo_va = amdgpu_vm_bo_find(vm, *bo); in kfd_queue_unref_bo_va() 372 if (bo_va && bo_va->queue_refcount) in kfd_queue_unref_bo_va() 373 bo_va->queue_refcount--; in kfd_queue_unref_bo_va()
|
| H A D | kfd_svm.c | 2893 bo = mapping->bo_va->base.bo; in svm_range_check_vm_userptr()
|