/linux/drivers/gpu/drm/xe/ |
H A D | xe_vm.h | 27 struct xe_vm *xe_vm_create(struct xe_device *xe, u32 flags); 29 struct xe_vm *xe_vm_lookup(struct xe_file *xef, u32 id); 32 static inline struct xe_vm *xe_vm_get(struct xe_vm *vm) in xe_vm_get() 38 static inline void xe_vm_put(struct xe_vm *vm) in xe_vm_put() 43 int xe_vm_lock(struct xe_vm *vm, bool intr); 45 void xe_vm_unlock(struct xe_vm *vm); 47 static inline bool xe_vm_is_closed(struct xe_vm *vm) in xe_vm_is_closed() 53 static inline bool xe_vm_is_banned(struct xe_vm *vm) in xe_vm_is_banned() 58 static inline bool xe_vm_is_closed_or_banned(struct xe_vm *vm) in xe_vm_is_closed_or_banned() 65 xe_vm_find_overlapping_vma(struct xe_vm *vm, u64 start, u64 range); [all …]
|
H A D | xe_trace_bo.h | 31 __field(struct xe_vm *, vm) 175 DECLARE_EVENT_CLASS(xe_vm, 176 TP_PROTO(struct xe_vm *vm), 181 __field(struct xe_vm *, vm) 195 DEFINE_EVENT(xe_vm, xe_vm_kill, 196 TP_PROTO(struct xe_vm *vm), 200 DEFINE_EVENT(xe_vm, xe_vm_create, 201 TP_PROTO(struct xe_vm *vm), 205 DEFINE_EVENT(xe_vm, xe_vm_free, 206 TP_PROTO(struct xe_vm *vm), [all …]
|
H A D | xe_exec_queue.h | 17 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, 22 struct xe_vm *vm, 76 void xe_exec_queue_last_fence_put(struct xe_exec_queue *e, struct xe_vm *vm); 79 struct xe_vm *vm); 81 struct xe_vm *vm); 82 void xe_exec_queue_last_fence_set(struct xe_exec_queue *e, struct xe_vm *vm, 85 struct xe_vm *vm);
|
H A D | xe_vm.c | 43 static struct drm_gem_object *xe_vm_obj(struct xe_vm *vm) in xe_vm_obj() 70 struct xe_vm *vm = xe_vma_vm(vma); in xe_vma_userptr_pin_pages() 79 static bool preempt_fences_waiting(struct xe_vm *vm) in preempt_fences_waiting() 105 static int alloc_preempt_fences(struct xe_vm *vm, struct list_head *list, in alloc_preempt_fences() 126 static int wait_for_existing_preempt_fences(struct xe_vm *vm) in wait_for_existing_preempt_fences() 148 static bool xe_vm_is_idle(struct xe_vm *vm) in xe_vm_is_idle() 161 static void arm_preempt_fences(struct xe_vm *vm, struct list_head *list) in arm_preempt_fences() 180 static int add_preempt_fences(struct xe_vm *vm, struct xe_bo *bo) in add_preempt_fences() 204 static void resume_and_reinstall_preempt_fences(struct xe_vm *vm, in resume_and_reinstall_preempt_fences() 220 int xe_vm_add_compute_exec_queue(struct xe_vm *vm, struct xe_exec_queue *q) in xe_vm_add_compute_exec_queue() [all …]
|
H A D | xe_bo.h | 78 struct xe_tile *tile, struct xe_vm *vm, 82 struct xe_vm *vm, size_t size, 85 struct xe_vm *vm, size_t size, 88 struct xe_vm *vm, size_t size, 92 struct xe_vm *vm, size_t size, 95 struct xe_vm *vm, size_t size, u64 offset, 99 struct xe_vm *vm, 168 int xe_bo_validate(struct xe_bo *bo, struct xe_vm *vm, bool allow_res_evict);
|
H A D | xe_pt.h | 18 struct xe_vm; 30 struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_tile *tile, 33 void xe_pt_populate_empty(struct xe_tile *tile, struct xe_vm *vm,
|
H A D | xe_exec_queue.c | 50 struct xe_vm *vm, in __xe_exec_queue_alloc() 111 struct xe_vm *vm = q->vm; in __xe_exec_queue_init() 146 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, in xe_exec_queue_create() 171 struct xe_vm *vm, in xe_exec_queue_create_class() 218 struct xe_vm *migrate_vm; in xe_exec_queue_create_bind() 541 struct xe_vm *vm; in xe_exec_queue_create_ioctl() 854 struct xe_vm *vm) in xe_exec_queue_last_fence_lockdep_assert() 869 void xe_exec_queue_last_fence_put(struct xe_exec_queue *q, struct xe_vm *vm) in xe_exec_queue_last_fence_put() 900 struct xe_vm *vm) in xe_exec_queue_last_fence_get() 927 struct xe_vm *vm) in xe_exec_queue_last_fence_get_for_resume() [all …]
|
H A D | xe_bo_types.h | 19 struct xe_vm; 35 struct xe_vm *vm;
|
H A D | xe_sync.h | 15 struct xe_vm; 32 struct xe_exec_queue *q, struct xe_vm *vm);
|
H A D | xe_vm_types.h | 23 struct xe_vm; 138 struct xe_vm { struct 363 struct xe_vm *vm;
|
H A D | xe_gt_pagefault.c | 83 static struct xe_vma *lookup_vma(struct xe_vm *vm, u64 page_addr) in lookup_vma() 101 struct xe_vm *vm = xe_vma_vm(vma); in xe_pf_begin() 131 struct xe_vm *vm = xe_vma_vm(vma); in handle_vma_pagefault() 188 static struct xe_vm *asid_to_vm(struct xe_device *xe, u32 asid) in asid_to_vm() 190 struct xe_vm *vm; in asid_to_vm() 207 struct xe_vm *vm; in handle_pagefault() 538 static struct xe_vma *get_acc_vma(struct xe_vm *vm, struct acc *acc) in get_acc_vma() 551 struct xe_vm *vm; in handle_acc()
|
H A D | xe_migrate.h | 23 struct xe_vm; 114 struct xe_vm *xe_migrate_get_vm(struct xe_migrate *m);
|
H A D | xe_sched_job.h | 12 struct xe_vm; 61 int xe_sched_job_last_fence_add_dep(struct xe_sched_job *job, struct xe_vm *vm);
|
H A D | xe_exec_queue_types.h | 22 struct xe_vm; 55 struct xe_vm *vm;
|
H A D | xe_pt.c | 56 static u64 __xe_pt_empty_pte(struct xe_tile *tile, struct xe_vm *vm, in __xe_pt_empty_pte() 97 struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_tile *tile, in xe_pt_create() 150 void xe_pt_populate_empty(struct xe_tile *tile, struct xe_vm *vm, in xe_pt_populate_empty() 258 struct xe_vm *vm; 486 struct xe_vm *vm = xe_walk->vm; in xe_pt_stage_bind_entry() 877 struct xe_vm *vm = xe_vma_vm(vma); in xe_pt_commit_locks_assert() 1066 static int op_add_deps(struct xe_vm *vm, struct xe_vma_op *op, in op_add_deps() 1097 struct xe_vm *vm, in xe_pt_vm_dependencies() 1170 struct xe_vm *vm = vops->vm; in xe_pt_pre_commit() 1203 static int vma_check_userptr(struct xe_vm *vm, struct xe_vma *vma, in vma_check_userptr() [all …]
|
H A D | xe_exec.c | 99 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn() 119 struct xe_vm *vm; in xe_exec_ioctl()
|
H A D | xe_hmm.c | 129 struct xe_vm *vm = xe_vma_vm(vma); in xe_hmm_userptr_free_sg() 174 struct xe_vm *vm = xe_vma_vm(vma); in xe_hmm_userptr_populate_range()
|
H A D | xe_bo.c | 505 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); in xe_bo_trigger_rebind() 1470 struct xe_tile *tile, struct xe_vm *vm, in __xe_bo_create_locked() 1542 struct xe_tile *tile, struct xe_vm *vm, in xe_bo_create_locked_range() 1551 struct xe_vm *vm, size_t size, in xe_bo_create_locked() 1559 struct xe_vm *vm, size_t size, in xe_bo_create_user() 1573 struct xe_vm *vm, size_t size, in xe_bo_create() 1585 struct xe_vm *vm, in xe_bo_create_pin_map_at() 1595 struct xe_vm *vm, in xe_bo_create_pin_map_at_aligned() 1636 struct xe_vm *vm, size_t size, in xe_bo_create_pin_map() 1906 int xe_bo_validate(struct xe_bo *bo, struct xe_vm *vm, bool allow_res_evict) in xe_bo_validate() [all …]
|
H A D | xe_sched_job.c | 244 struct xe_vm *vm = q->vm; in xe_sched_job_arm() 303 int xe_sched_job_last_fence_add_dep(struct xe_sched_job *job, struct xe_vm *vm) in xe_sched_job_last_fence_add_dep()
|
H A D | xe_lrc.c | 857 static void xe_lrc_set_ppgtt(struct xe_lrc *lrc, struct xe_vm *vm) in xe_lrc_set_ppgtt() 878 struct xe_vm *vm, u32 ring_size) in xe_lrc_init() 1013 struct xe_lrc *xe_lrc_create(struct xe_hw_engine *hwe, struct xe_vm *vm, in xe_lrc_create() 1656 struct xe_vm *vm; in xe_lrc_snapshot_capture_delayed() 1735 struct xe_vm *vm; in xe_lrc_snapshot_free()
|
H A D | xe_sync.c | 281 struct xe_exec_queue *q, struct xe_vm *vm) in xe_sync_in_fence_get()
|
H A D | Makefile | 110 xe_vm.o \
|
H A D | xe_device.c | 148 struct xe_vm *vm; in xe_file_close()
|
/linux/drivers/gpu/drm/xe/tests/ |
H A D | xe_bo.c | 205 struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate); in evict_test_run_tile()
|