| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_svm.h | 26 struct xe_vm; 82 int xe_svm_init(struct xe_vm *vm); 84 void xe_svm_fini(struct xe_vm *vm); 86 void xe_svm_close(struct xe_vm *vm); 88 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, 92 bool xe_svm_has_mapping(struct xe_vm *vm, u64 start, u64 end); 101 struct xe_svm_range *xe_svm_range_find_or_insert(struct xe_vm *vm, u64 addr, 104 int xe_svm_range_get_pages(struct xe_vm *vm, struct xe_svm_range *range, 110 void xe_svm_range_migrate_to_smem(struct xe_vm *vm, struct xe_svm_range *range); 112 bool xe_svm_range_validate(struct xe_vm *vm, [all …]
|
| H A D | xe_trace_bo.h | 31 __field(struct xe_vm *, vm) 95 __field(struct xe_vm *, vm) 188 DECLARE_EVENT_CLASS(xe_vm, 189 TP_PROTO(struct xe_vm *vm), 194 __field(struct xe_vm *, vm) 211 DEFINE_EVENT(xe_vm, xe_vm_kill, 212 TP_PROTO(struct xe_vm *vm), 216 DEFINE_EVENT(xe_vm, xe_vm_create, 217 TP_PROTO(struct xe_vm *vm), 221 DEFINE_EVENT(xe_vm, xe_vm_free, [all …]
|
| H A D | xe_vm.c | 46 static struct drm_gem_object *xe_vm_obj(struct xe_vm *vm) in xe_vm_obj() 60 int xe_vm_drm_exec_lock(struct xe_vm *vm, struct drm_exec *exec) in xe_vm_drm_exec_lock() 65 static bool preempt_fences_waiting(struct xe_vm *vm) in preempt_fences_waiting() 91 static int alloc_preempt_fences(struct xe_vm *vm, struct list_head *list, in alloc_preempt_fences() 112 static int wait_for_existing_preempt_fences(struct xe_vm *vm) in wait_for_existing_preempt_fences() 144 static bool xe_vm_is_idle(struct xe_vm *vm) in xe_vm_is_idle() 157 static void arm_preempt_fences(struct xe_vm *vm, struct list_head *list) in arm_preempt_fences() 176 static int add_preempt_fences(struct xe_vm *vm, struct xe_bo *bo) in add_preempt_fences() 200 static void resume_and_reinstall_preempt_fences(struct xe_vm *vm, in resume_and_reinstall_preempt_fences() 216 int xe_vm_add_compute_exec_queue(struct xe_vm *vm, struct xe_exec_queue *q) in xe_vm_add_compute_exec_queue() [all …]
|
| H A D | xe_svm.c | 53 static int xe_svm_get_pagemaps(struct xe_vm *vm); 55 void *xe_svm_private_page_owner(struct xe_vm *vm, bool force_smem) in xe_svm_private_page_owner() 81 static struct xe_vm *gpusvm_to_vm(struct drm_gpusvm *gpusvm) in gpusvm_to_vm() 83 return container_of(gpusvm, struct xe_vm, svm.gpusvm); in gpusvm_to_vm() 86 static struct xe_vm *range_to_vm(struct drm_gpusvm_range *r) in range_to_vm() 130 xe_svm_garbage_collector_add_range(struct xe_vm *vm, struct xe_svm_range *range, in xe_svm_garbage_collector_add_range() 154 xe_svm_range_notifier_event_begin(struct xe_vm *vm, struct drm_gpusvm_range *r, in xe_svm_range_notifier_event_begin() 204 xe_svm_range_notifier_event_end(struct xe_vm *vm, struct drm_gpusvm_range *r, in xe_svm_range_notifier_event_end() 228 struct xe_vm *vm = gpusvm_to_vm(gpusvm); in xe_svm_invalidate() 299 static int __xe_svm_garbage_collector(struct xe_vm *vm, in __xe_svm_garbage_collector() [all …]
|
| H A D | xe_vm_madvise.c | 44 static int get_vmas(struct xe_vm *vm, struct xe_vmas_in_madvise_range *madvise_range) in get_vmas() 94 static void madvise_preferred_mem_loc(struct xe_device *xe, struct xe_vm *vm, in madvise_preferred_mem_loc() 127 static void madvise_atomic(struct xe_device *xe, struct xe_vm *vm, in madvise_atomic() 169 static void madvise_pat_index(struct xe_device *xe, struct xe_vm *vm, in madvise_pat_index() 341 static void madvise_purgeable(struct xe_device *xe, struct xe_vm *vm, in madvise_purgeable() 404 typedef void (*madvise_func)(struct xe_device *xe, struct xe_vm *vm, 416 static u8 xe_zap_ptes_in_madvise_range(struct xe_vm *vm, u64 start, u64 end) in xe_zap_ptes_in_madvise_range() 458 static int xe_vm_invalidate_madvise_range(struct xe_vm *vm, u64 start, u64 end) in xe_vm_invalidate_madvise_range() 568 static int xe_madvise_details_init(struct xe_vm *vm, const struct drm_xe_madvise *args, in xe_madvise_details_init() 663 static bool check_bo_args_are_sane(struct xe_vm *vm, struct xe_vma **vmas, in check_bo_args_are_sane() [all …]
|
| H A D | xe_bo_types.h | 22 struct xe_vm; 42 struct xe_vm *vm;
|
| H A D | xe_pagefault.c | 52 struct xe_vm *vm = xe_vma_vm(vma); in xe_pagefault_begin() 82 struct xe_vm *vm = xe_vma_vm(vma); in xe_pagefault_handle_vma() 155 static struct xe_vm *xe_pagefault_asid_to_vm(struct xe_device *xe, u32 asid) in xe_pagefault_asid_to_vm() 157 struct xe_vm *vm; in xe_pagefault_asid_to_vm() 174 struct xe_vm *vm; in xe_pagefault_service() 268 struct xe_vm *vm; in xe_pagefault_save_to_vm()
|
| H A D | xe_pt.c | 61 static u64 __xe_pt_empty_pte(struct xe_tile *tile, struct xe_vm *vm, in __xe_pt_empty_pte() 103 struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_tile *tile, in xe_pt_create() 163 void xe_pt_populate_empty(struct xe_tile *tile, struct xe_vm *vm, in xe_pt_populate_empty() 288 struct xe_vm *vm; 525 struct xe_vm *vm = xe_walk->vm; in xe_pt_stage_bind_entry() 652 static bool xe_atomic_for_vram(struct xe_vm *vm, struct xe_vma *vma) in xe_atomic_for_vram() 660 static bool xe_atomic_for_system(struct xe_vm *vm, struct xe_vma *vma) in xe_atomic_for_system() 711 struct xe_vm *vm = xe_vma_vm(vma); in xe_pt_stage_bind() 971 bool xe_pt_zap_ptes_range(struct xe_tile *tile, struct xe_vm *vm, in xe_pt_zap_ptes_range() 1055 struct xe_vm *vm; in xe_pt_commit_prepare_locks_assert() [all …]
|
| H A D | xe_vm_types.h | 31 struct xe_vm; 209 struct xe_vm { struct 533 struct xe_vm *vm;
|
| H A D | xe_exec_queue.c | 201 struct xe_vm *vm, in __xe_exec_queue_alloc() 419 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, in xe_exec_queue_create() 476 struct xe_vm *vm, in xe_exec_queue_create_class() 520 struct xe_vm *user_vm, in xe_exec_queue_create_bind() 525 struct xe_vm *migrate_vm; in xe_exec_queue_create_bind() 1269 struct xe_vm *vm; in xe_exec_queue_create_ioctl() 1629 struct xe_vm *vm) in xe_exec_queue_last_fence_lockdep_assert() 1646 void xe_exec_queue_last_fence_put(struct xe_exec_queue *q, struct xe_vm *vm) in xe_exec_queue_last_fence_put() 1677 struct xe_vm *vm) in xe_exec_queue_last_fence_get() 1704 struct xe_vm *vm) in xe_exec_queue_last_fence_get_for_resume() [all …]
|
| H A D | xe_tlb_inval_job.c | 28 struct xe_vm *vm; 100 struct xe_vm *vm, u64 start, u64 end, int type) in xe_tlb_inval_job_create() 182 struct xe_vm *vm = job->vm; in xe_tlb_inval_job_destroy()
|
| H A D | xe_exec.c | 101 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn() 126 struct xe_vm *vm; in xe_exec_ioctl()
|
| H A D | xe_lrc.h | 21 struct xe_vm; 54 struct xe_lrc *xe_lrc_create(struct xe_hw_engine *hwe, struct xe_vm *vm,
|
| H A D | xe_migrate.c | 143 static void xe_migrate_program_identity(struct xe_device *xe, struct xe_vm *vm, struct xe_bo *bo, in xe_migrate_program_identity() 189 struct xe_vm *vm, struct drm_exec *exec) in xe_migrate_pt_bo_alloc() 217 struct xe_vm *vm, u32 *ofs) in xe_migrate_prepare_vm() 428 static int xe_migrate_lock_prepare_vm(struct xe_tile *tile, struct xe_migrate *m, struct xe_vm *vm) in xe_migrate_lock_prepare_vm() 467 struct xe_vm *vm; in xe_migrate_init() 632 struct xe_vm *vm = m->q->vm; in emit_pte() 1767 struct xe_vm *xe_migrate_get_vm(struct xe_migrate *m) in xe_migrate_get_vm() 1790 struct xe_vm *vm = pt_update->vops->vm; in xe_migrate_update_pgtables_cpu() 1926 struct xe_vm *vm = pt_update->vops->vm; in __xe_migrate_update_pgtables()
|
| H A D | xe_device.c | 167 struct xe_vm *vm; in xe_file_close() 1419 struct xe_vm *xe_device_asid_to_vm(struct xe_device *xe, u32 asid) in xe_device_asid_to_vm() 1421 struct xe_vm *vm; in xe_device_asid_to_vm()
|
| H A D | xe_bo.c | 685 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); in xe_bo_trigger_rebind() 2482 struct xe_tile *tile, struct xe_vm *vm, in __xe_bo_create_locked() 2580 struct xe_vm *vm, size_t size, in xe_bo_create_locked() 2629 struct xe_vm *vm, size_t size, in xe_bo_create_user() 2703 struct xe_vm *vm, in xe_bo_create_pin_map_at_aligned() 2807 struct xe_vm *vm, size_t size, in xe_bo_create_pin_map() 3115 int xe_bo_validate(struct xe_bo *bo, struct xe_vm *vm, bool allow_res_evict, in xe_bo_validate() 3336 struct xe_vm *vm = NULL; in xe_gem_create_ioctl()
|
| H A D | xe_sync.c | 333 struct xe_exec_queue *q, struct xe_vm *vm) in xe_sync_in_fence_get()
|
| H A D | xe_lrc.c | 1041 static void xe_lrc_set_ppgtt(struct xe_lrc *lrc, struct xe_vm *vm) in xe_lrc_set_ppgtt() 1472 static int xe_lrc_ctx_init(struct xe_lrc *lrc, struct xe_hw_engine *hwe, struct xe_vm *vm, in xe_lrc_ctx_init() 1603 static int xe_lrc_init(struct xe_lrc *lrc, struct xe_hw_engine *hwe, struct xe_vm *vm, in xe_lrc_init() 1687 struct xe_lrc *xe_lrc_create(struct xe_hw_engine *hwe, struct xe_vm *vm, in xe_lrc_create()
|
| H A D | xe_pm.c | 385 struct xe_vm *vm, *next; in xe_pm_wake_rebind_workers()
|
| H A D | Makefile | 142 xe_vm.o \
|
| /linux/drivers/gpu/drm/xe/tests/ |
| H A D | xe_bo.c | 211 struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate); in evict_test_run_tile()
|