Home
last modified time | relevance | path

Searched refs:xe_vma (Results 1 – 9 of 9) sorted by relevance

/linux/drivers/gpu/drm/xe/
H A Dxe_trace_bo.h88 DECLARE_EVENT_CLASS(xe_vma,
89 TP_PROTO(struct xe_vma *vma),
94 __field(struct xe_vma *, vma)
118 DEFINE_EVENT(xe_vma, xe_vma_flush,
119 TP_PROTO(struct xe_vma *vma),
123 DEFINE_EVENT(xe_vma, xe_vma_pagefault,
124 TP_PROTO(struct xe_vma *vma),
128 DEFINE_EVENT(xe_vma, xe_vma_acc,
129 TP_PROTO(struct xe_vma *vma),
133 DEFINE_EVENT(xe_vma, xe_vma_bind,
[all …]
H A Dxe_svm.h27 struct xe_vma;
88 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma,
102 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx);
107 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma,
116 u64 xe_svm_find_vma_start(struct xe_vm *vm, u64 addr, u64 end, struct xe_vma *vma);
122 struct drm_pagemap *xe_vma_resolve_pagemap(struct xe_vma *vma, struct xe_tile *tile);
203 struct xe_vma;
258 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, in xe_svm_handle_pagefault()
291 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx) in xe_svm_range_find_or_insert()
324 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma, in xe_svm_range_needs_migrate_to_vram()
[all …]
H A Dxe_vm_types.h112 struct xe_vma { struct
185 struct xe_vma vma; argument
357 struct xe_vma *last_fault_vma;
414 struct xe_vma *vma;
429 struct xe_vma *prev;
431 struct xe_vma *next;
457 struct xe_vma *vma;
H A Dxe_vm_madvise.c21 struct xe_vma **vmas;
49 struct xe_vma **__vmas; in get_vmas()
63 struct xe_vma *vma = gpuva_to_vma(gpuva); in get_vmas()
95 struct xe_vma **vmas, int num_vmas, in madvise_preferred_mem_loc()
104 struct xe_vma *vma = vmas[i]; in madvise_preferred_mem_loc()
128 struct xe_vma **vmas, int num_vmas, in madvise_atomic()
170 struct xe_vma **vmas, int num_vmas, in madvise_pat_index()
270 struct xe_vma *vma = gpuva_to_vma(gpuva); in xe_bo_all_vmas_dontneed()
342 struct xe_vma **vmas, int num_vmas, in madvise_purgeable()
405 struct xe_vma **vmas, int num_vmas,
[all …]
H A Dxe_vm.c679 struct xe_vma *vma; in xe_vma_svm_prefetch_op_fini()
727 static void xe_vm_populate_rebind(struct xe_vma_op *op, struct xe_vma *vma, in xe_vm_populate_rebind()
742 static int xe_vm_ops_add_rebind(struct xe_vma_ops *vops, struct xe_vma *vma, in xe_vm_ops_add_rebind()
767 struct xe_vma *vma, *next; in xe_vm_rebind()
819 struct dma_fence *xe_vma_rebind(struct xe_vm *vm, struct xe_vma *vma, u8 tile_mask) in xe_vma_rebind()
863 struct xe_vma *vma, in xe_vm_populate_range_rebind()
877 struct xe_vma *vma, in xe_vm_ops_add_range_rebind()
907 struct xe_vma *vma, in xe_vm_range_rebind()
1041 static void xe_vma_free(struct xe_vma *vma) in xe_vma_free()
1067 static struct xe_vma *xe_vma_create(struct xe_vm *vm, in xe_vma_create()
[all …]
H A Dxe_pt.c307 struct xe_vma *vma;
652 static bool xe_atomic_for_vram(struct xe_vm *vm, struct xe_vma *vma) in xe_atomic_for_vram()
660 static bool xe_atomic_for_system(struct xe_vm *vm, struct xe_vma *vma) in xe_atomic_for_system()
703 xe_pt_stage_bind(struct xe_tile *tile, struct xe_vma *vma, in xe_pt_stage_bind()
929 bool xe_pt_zap_ptes(struct xe_tile *tile, struct xe_vma *vma) in xe_pt_zap_ptes()
1027 static void xe_pt_cancel_bind(struct xe_vma *vma, in xe_pt_cancel_bind()
1051 #define XE_INVALID_VMA ((struct xe_vma *)(0xdeaddeadull))
1053 static void xe_pt_commit_prepare_locks_assert(struct xe_vma *vma) in xe_pt_commit_prepare_locks_assert()
1069 static void xe_pt_commit_locks_assert(struct xe_vma *vma) in xe_pt_commit_locks_assert()
1083 static void xe_pt_commit(struct xe_vma *vma, in xe_pt_commit()
[all …]
H A Dxe_pagefault.c48 static int xe_pagefault_begin(struct drm_exec *exec, struct xe_vma *vma, in xe_pagefault_begin()
79 static int xe_pagefault_handle_vma(struct xe_gt *gt, struct xe_vma *vma, in xe_pagefault_handle_vma()
175 struct xe_vma *vma = NULL; in xe_pagefault_service()
H A Dxe_svm.c318 static void xe_vma_set_default_attributes(struct xe_vma *vma) in xe_vma_set_default_attributes()
333 struct xe_vma *vma; in xe_svm_range_set_default_attr()
1042 u64 xe_svm_find_vma_start(struct xe_vm *vm, u64 start, u64 end, struct xe_vma *vma) in xe_svm_find_vma_start()
1140 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma, in xe_svm_range_needs_migrate_to_vram()
1213 static int __xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma,
1386 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, in xe_svm_handle_pagefault()
1489 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx) in xe_svm_range_find_or_insert()
1595 struct drm_pagemap *xe_vma_resolve_pagemap(struct xe_vma *vma, struct xe_tile *tile) in xe_vma_resolve_pagemap()
2034 struct drm_pagemap *xe_vma_resolve_pagemap(struct xe_vma *vma, struct xe_tile *tile) in xe_vma_resolve_pagemap()
H A Dxe_bo.c719 struct xe_vma *vma = gpuva_to_vma(gpuva); in xe_bo_trigger_rebind()