| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_trace_bo.h | 88 DECLARE_EVENT_CLASS(xe_vma, 89 TP_PROTO(struct xe_vma *vma), 94 __field(struct xe_vma *, vma) 118 DEFINE_EVENT(xe_vma, xe_vma_flush, 119 TP_PROTO(struct xe_vma *vma), 123 DEFINE_EVENT(xe_vma, xe_vma_pagefault, 124 TP_PROTO(struct xe_vma *vma), 128 DEFINE_EVENT(xe_vma, xe_vma_acc, 129 TP_PROTO(struct xe_vma *vma), 133 DEFINE_EVENT(xe_vma, xe_vma_bind, [all …]
|
| H A D | xe_svm.h | 27 struct xe_vma; 88 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, 102 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx); 107 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma, 116 u64 xe_svm_find_vma_start(struct xe_vm *vm, u64 addr, u64 end, struct xe_vma *vma); 122 struct drm_pagemap *xe_vma_resolve_pagemap(struct xe_vma *vma, struct xe_tile *tile); 203 struct xe_vma; 258 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, in xe_svm_handle_pagefault() 291 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx) in xe_svm_range_find_or_insert() 324 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma, in xe_svm_range_needs_migrate_to_vram() [all …]
|
| H A D | xe_vm_types.h | 112 struct xe_vma { struct 185 struct xe_vma vma; argument 357 struct xe_vma *last_fault_vma; 414 struct xe_vma *vma; 429 struct xe_vma *prev; 431 struct xe_vma *next; 457 struct xe_vma *vma;
|
| H A D | xe_vm_madvise.c | 21 struct xe_vma **vmas; 49 struct xe_vma **__vmas; in get_vmas() 63 struct xe_vma *vma = gpuva_to_vma(gpuva); in get_vmas() 95 struct xe_vma **vmas, int num_vmas, in madvise_preferred_mem_loc() 104 struct xe_vma *vma = vmas[i]; in madvise_preferred_mem_loc() 128 struct xe_vma **vmas, int num_vmas, in madvise_atomic() 170 struct xe_vma **vmas, int num_vmas, in madvise_pat_index() 270 struct xe_vma *vma = gpuva_to_vma(gpuva); in xe_bo_all_vmas_dontneed() 342 struct xe_vma **vmas, int num_vmas, in madvise_purgeable() 405 struct xe_vma **vmas, int num_vmas, [all …]
|
| H A D | xe_vm.c | 679 struct xe_vma *vma; in xe_vma_svm_prefetch_op_fini() 727 static void xe_vm_populate_rebind(struct xe_vma_op *op, struct xe_vma *vma, in xe_vm_populate_rebind() 742 static int xe_vm_ops_add_rebind(struct xe_vma_ops *vops, struct xe_vma *vma, in xe_vm_ops_add_rebind() 767 struct xe_vma *vma, *next; in xe_vm_rebind() 819 struct dma_fence *xe_vma_rebind(struct xe_vm *vm, struct xe_vma *vma, u8 tile_mask) in xe_vma_rebind() 863 struct xe_vma *vma, in xe_vm_populate_range_rebind() 877 struct xe_vma *vma, in xe_vm_ops_add_range_rebind() 907 struct xe_vma *vma, in xe_vm_range_rebind() 1041 static void xe_vma_free(struct xe_vma *vma) in xe_vma_free() 1067 static struct xe_vma *xe_vma_create(struct xe_vm *vm, in xe_vma_create() [all …]
|
| H A D | xe_pt.c | 307 struct xe_vma *vma; 652 static bool xe_atomic_for_vram(struct xe_vm *vm, struct xe_vma *vma) in xe_atomic_for_vram() 660 static bool xe_atomic_for_system(struct xe_vm *vm, struct xe_vma *vma) in xe_atomic_for_system() 703 xe_pt_stage_bind(struct xe_tile *tile, struct xe_vma *vma, in xe_pt_stage_bind() 929 bool xe_pt_zap_ptes(struct xe_tile *tile, struct xe_vma *vma) in xe_pt_zap_ptes() 1027 static void xe_pt_cancel_bind(struct xe_vma *vma, in xe_pt_cancel_bind() 1051 #define XE_INVALID_VMA ((struct xe_vma *)(0xdeaddeadull)) 1053 static void xe_pt_commit_prepare_locks_assert(struct xe_vma *vma) in xe_pt_commit_prepare_locks_assert() 1069 static void xe_pt_commit_locks_assert(struct xe_vma *vma) in xe_pt_commit_locks_assert() 1083 static void xe_pt_commit(struct xe_vma *vma, in xe_pt_commit() [all …]
|
| H A D | xe_pagefault.c | 48 static int xe_pagefault_begin(struct drm_exec *exec, struct xe_vma *vma, in xe_pagefault_begin() 79 static int xe_pagefault_handle_vma(struct xe_gt *gt, struct xe_vma *vma, in xe_pagefault_handle_vma() 175 struct xe_vma *vma = NULL; in xe_pagefault_service()
|
| H A D | xe_svm.c | 318 static void xe_vma_set_default_attributes(struct xe_vma *vma) in xe_vma_set_default_attributes() 333 struct xe_vma *vma; in xe_svm_range_set_default_attr() 1042 u64 xe_svm_find_vma_start(struct xe_vm *vm, u64 start, u64 end, struct xe_vma *vma) in xe_svm_find_vma_start() 1140 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma, in xe_svm_range_needs_migrate_to_vram() 1213 static int __xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, 1386 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, in xe_svm_handle_pagefault() 1489 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx) in xe_svm_range_find_or_insert() 1595 struct drm_pagemap *xe_vma_resolve_pagemap(struct xe_vma *vma, struct xe_tile *tile) in xe_vma_resolve_pagemap() 2034 struct drm_pagemap *xe_vma_resolve_pagemap(struct xe_vma *vma, struct xe_tile *tile) in xe_vma_resolve_pagemap()
|
| H A D | xe_bo.c | 719 struct xe_vma *vma = gpuva_to_vma(gpuva); in xe_bo_trigger_rebind()
|