| /linux/drivers/gpu/drm/ |
| H A D | drm_gpuvm.c | 829 __restore_vm_bo_list(struct drm_gpuvm *gpuvm, spinlock_t *lock, in __restore_vm_bo_list() argument 871 __drm_gpuvm_bo_list_add(struct drm_gpuvm *gpuvm, spinlock_t *lock, in __drm_gpuvm_bo_list_add() argument 921 __drm_gpuvm_bo_list_del(struct drm_gpuvm *gpuvm, spinlock_t *lock, in __drm_gpuvm_bo_list_del() argument 976 static int __drm_gpuva_insert(struct drm_gpuvm *gpuvm, 989 drm_gpuvm_warn_check_overflow(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_warn_check_overflow() argument 991 return drm_WARN(gpuvm->drm, drm_gpuvm_check_overflow(addr, range), in drm_gpuvm_warn_check_overflow() 996 drm_gpuvm_in_mm_range(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_in_mm_range() argument 999 u64 mm_start = gpuvm->mm_start; in drm_gpuvm_in_mm_range() 1000 u64 mm_end = mm_start + gpuvm->mm_range; in drm_gpuvm_in_mm_range() 1006 drm_gpuvm_in_kernel_node(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_in_kernel_node() argument [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_vm.h | 36 drm_gpuvm_get(&vm->gpuvm); in xe_vm_get() 42 drm_gpuvm_put(&vm->gpuvm); in xe_vm_put() 91 static inline struct xe_vm *gpuvm_to_vm(struct drm_gpuvm *gpuvm) in gpuvm_to_vm() argument 93 return container_of(gpuvm, struct xe_vm, gpuvm); in gpuvm_to_vm() 143 return container_of(vma->gpuva.vm, struct xe_vm, gpuvm); in xe_vma_vm() 285 return drm_gpuvm_resv(&vm->gpuvm); in xe_vm_resv()
|
| H A D | xe_vm.c | 46 return vm->gpuvm.r_obj; in xe_vm_obj() 209 drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, q->lr.pfence, in resume_and_reinstall_preempt_fences() 217 .vm = &vm->gpuvm, in xe_vm_add_compute_exec_queue() 247 drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, pfence, in xe_vm_add_compute_exec_queue() 371 ret = drm_gpuvm_validate(&vm->gpuvm, exec); in xe_vm_validate_rebind() 378 } while (!list_empty(&vm->gpuvm.evict.list)); in xe_vm_validate_rebind() 394 err = drm_gpuvm_prepare_vm(&vm->gpuvm, exec, 0); in xe_preempt_work_begin() 409 err = drm_gpuvm_prepare_objects(&vm->gpuvm, exec, 0); in xe_preempt_work_begin() 1029 vma->gpuva.vm = &vm->gpuvm; in xe_vma_create() 1202 gpuva = drm_gpuva_find_first(&vm->gpuvm, start, range); in xe_vm_find_overlapping_vma() [all …]
|
| H A D | xe_exec.c | 101 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn() 269 vm_exec.vm = &vm->gpuvm; in xe_exec_ioctl() 325 drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, &job->drm.s_fence->finished, in xe_exec_ioctl()
|
| H A D | xe_vm_types.h | 180 struct drm_gpuvm gpuvm; member
|
| H A D | xe_vm_madvise.c | 56 drm_gpuvm_for_each_va_range(gpuva, &vm->gpuvm, addr, addr + range) { in get_vmas() 206 drm_gpuvm_for_each_va_range(gpuva, &vm->gpuvm, start, end) { in xe_zap_ptes_in_madvise_range()
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_gem_vma.c | 159 msm_gem_vm_free(struct drm_gpuvm *gpuvm) in msm_gem_vm_free() argument 161 struct msm_gem_vm *vm = container_of(gpuvm, struct msm_gem_vm, base); in msm_gem_vm_free() 177 msm_gem_vm_unusable(struct drm_gpuvm *gpuvm) in msm_gem_vm_unusable() argument 179 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vm_unusable() 368 msm_gem_vma_new(struct drm_gpuvm *gpuvm, struct drm_gem_object *obj, in msm_gem_vma_new() argument 371 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vma_new() 895 msm_gem_vm_close(struct drm_gpuvm *gpuvm) in msm_gem_vm_close() argument 897 struct msm_gem_vm *vm = to_msm_vm(gpuvm); in msm_gem_vm_close() 918 drm_exec_lock_obj(&exec, drm_gpuvm_resv_obj(gpuvm)); in msm_gem_vm_close() 921 drm_gpuvm_for_each_va_safe (vma, tmp, gpuvm) { in msm_gem_vm_close() [all …]
|
| /linux/drivers/gpu/drm/imagination/ |
| H A D | pvr_vm.c | 70 struct pvr_vm_context *to_pvr_vm_context(struct drm_gpuvm *gpuvm) in to_pvr_vm_context() argument 72 return container_of(gpuvm, struct pvr_vm_context, gpuvm_mgr); in to_pvr_vm_context() 507 static void pvr_gpuvm_free(struct drm_gpuvm *gpuvm) in pvr_gpuvm_free() argument 509 kfree(to_pvr_vm_context(gpuvm)); in pvr_gpuvm_free()
|
| /linux/drivers/gpu/drm/amd/display/dc/dml/ |
| H A D | display_mode_lib.c | 188 dml_print("DML PARAMS: gpuvm = %d\n", pipe_src->gpuvm); in dml_log_pipe_params()
|
| H A D | display_mode_structs.h | 392 bool gpuvm; // gpuvm enabled member
|
| H A D | display_mode_vba.c | 882 …mode_lib->vba.GPUVMEnable = mode_lib->vba.GPUVMEnable || !!pipes[k].pipe.src.gpuvm || !!pipes[k].p… in fetch_pipe_params()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_uvmm.c | 1865 nouveau_uvmm_free(struct drm_gpuvm *gpuvm) in nouveau_uvmm_free() argument 1867 struct nouveau_uvmm *uvmm = uvmm_from_gpuvm(gpuvm); in nouveau_uvmm_free()
|
| /linux/drivers/gpu/drm/amd/display/dc/resource/dcn31/ |
| H A D | dcn31_resource.c | 1628 pipes[i].pipe.src.gpuvm = 1; in dcn31x_populate_dml_pipes_from_context() 1673 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn31_populate_dml_pipes_from_context()
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_mmu.c | 1871 static void panthor_vm_free(struct drm_gpuvm *gpuvm) in panthor_vm_free() argument 1873 struct panthor_vm *vm = container_of(gpuvm, struct panthor_vm, base); in panthor_vm_free()
|
| /linux/drivers/gpu/drm/amd/display/dc/resource/dcn32/ |
| H A D | dcn32_resource.c | 1921 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn32_populate_dml_pipes_from_context()
|