Home
last modified time | relevance | path

Searched defs:vm (Results 1 – 25 of 222) sorted by relevance

123456789

/linux/tools/testing/selftests/kvm/s390/
H A Dcmma_test.c97 static void create_main_memslot(struct kvm_vm *vm) in create_main_memslot()
107 static void create_test_memslot(struct kvm_vm *vm) in create_test_memslot()
119 static void create_memslots(struct kvm_vm *vm) in create_memslots()
139 static void finish_vm_setup(struct kvm_vm *vm) in finish_vm_setup()
153 struct kvm_vm *vm; in create_vm_two_memslots() local
164 static void enable_cmma(struct kvm_vm *vm) in enable_cmma()
172 static void enable_dirty_tracking(struct kvm_vm *vm) in enable_dirty_tracking()
178 static int __enable_migration_mode(struct kvm_vm *vm) in __enable_migration_mode()
187 static void enable_migration_mode(struct kvm_vm *vm) in enable_migration_mode()
194 static bool is_migration_mode_on(struct kvm_vm *vm) in is_migration_mode_on()
[all …]
/linux/drivers/gpu/drm/xe/
H A Dxe_vm.h37 return vm; in xe_vm_put() argument
52 return !vm argument
31 xe_vm_get(struct xe_vm * vm) xe_vm_get() argument
46 xe_vm_is_closed(struct xe_vm * vm) xe_vm_is_closed() argument
57 xe_vm_is_closed_or_banned(struct xe_vm * vm) xe_vm_is_closed_or_banned() argument
72 xe_vm_has_scratch(const struct xe_vm * vm) xe_vm_has_scratch() argument
186 xe_vm_in_fault_mode(struct xe_vm * vm) xe_vm_in_fault_mode() argument
191 xe_vm_in_lr_mode(struct xe_vm * vm) xe_vm_in_lr_mode() argument
196 xe_vm_in_preempt_fence_mode(struct xe_vm * vm) xe_vm_in_preempt_fence_mode() argument
216 xe_vm_queue_rebind_worker(struct xe_vm * vm) xe_vm_queue_rebind_worker() argument
231 xe_vm_reactivate_rebind(struct xe_vm * vm) xe_vm_reactivate_rebind() argument
256 xe_vm_resv(struct xe_vm * vm) xe_vm_resv() argument
267 xe_vm_assert_held(vm) global() argument
[all...]
H A Dxe_vm.c42 xe_vm_obj(struct xe_vm * vm) xe_vm_obj() argument
69 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_userptr_pin_pages() local
78 preempt_fences_waiting(struct xe_vm * vm) preempt_fences_waiting() argument
104 alloc_preempt_fences(struct xe_vm * vm,struct list_head * list,unsigned int * count) alloc_preempt_fences() argument
125 wait_for_existing_preempt_fences(struct xe_vm * vm) wait_for_existing_preempt_fences() argument
147 xe_vm_is_idle(struct xe_vm * vm) xe_vm_is_idle() argument
160 arm_preempt_fences(struct xe_vm * vm,struct list_head * list) arm_preempt_fences() argument
179 add_preempt_fences(struct xe_vm * vm,struct xe_bo * bo) add_preempt_fences() argument
203 resume_and_reinstall_preempt_fences(struct xe_vm * vm,struct drm_exec * exec) resume_and_reinstall_preempt_fences() argument
219 xe_vm_add_compute_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) xe_vm_add_compute_exec_queue() argument
280 xe_vm_remove_compute_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) xe_vm_remove_compute_exec_queue() argument
309 __xe_vm_userptr_needs_repin(struct xe_vm * vm) __xe_vm_userptr_needs_repin() argument
327 xe_vm_kill(struct xe_vm * vm,bool unlocked) xe_vm_kill() argument
383 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); xe_gpuvm_validate() local
415 xe_vm_validate_rebind(struct xe_vm * vm,struct drm_exec * exec,unsigned int num_fences) xe_vm_validate_rebind() argument
441 xe_preempt_work_begin(struct drm_exec * exec,struct xe_vm * vm,bool * done) xe_preempt_work_begin() argument
480 struct xe_vm *vm = container_of(w, struct xe_vm, preempt.rebind_work); preempt_rebind_work_func() local
582 __vma_userptr_invalidate(struct xe_vm * vm,struct xe_userptr_vma * uvma) __vma_userptr_invalidate() argument
633 struct xe_vm *vm = xe_vma_vm(vma); vma_userptr_invalidate() local
668 struct xe_vm *vm = xe_vma_vm(&uvma->vma); xe_vma_userptr_force_invalidate() local
687 xe_vm_userptr_pin(struct xe_vm * vm) xe_vm_userptr_pin() argument
772 xe_vm_userptr_check_repin(struct xe_vm * vm) xe_vm_userptr_check_repin() argument
853 xe_vm_rebind(struct xe_vm * vm,bool rebind_worker) xe_vm_rebind() argument
908 xe_vma_rebind(struct xe_vm * vm,struct xe_vma * vma,u8 tile_mask) xe_vma_rebind() argument
962 xe_vma_create(struct xe_vm * vm,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 start,u64 end,u16 pat_index,unsigned int flags) xe_vma_create() argument
1066 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_destroy_late() local
1116 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_destroy() local
1160 struct xe_vm *vm = xe_vma_vm(vma); xe_vm_lock_vma() local
1192 xe_vm_find_overlapping_vma(struct xe_vm * vm,u64 start,u64 range) xe_vm_find_overlapping_vma() argument
1208 xe_vm_insert_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_insert_vma() argument
1223 xe_vm_remove_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_remove_vma() argument
1397 xe_vm_create_scratch(struct xe_device * xe,struct xe_tile * tile,struct xe_vm * vm) xe_vm_create_scratch() argument
1414 xe_vm_free_scratch(struct xe_vm * vm) xe_vm_free_scratch() argument
1437 struct xe_vm *vm; xe_vm_create() local
1583 xe_vm_close(struct xe_vm * vm) xe_vm_close() argument
1590 xe_vm_close_and_put(struct xe_vm * vm) xe_vm_close_and_put() argument
1694 struct xe_vm *vm = vm_destroy_work_func() local
1726 struct xe_vm *vm = container_of(gpuvm, struct xe_vm, gpuvm); xe_vm_free() local
1734 struct xe_vm *vm; xe_vm_lookup() local
1745 xe_vm_pdp4_descriptor(struct xe_vm * vm,struct xe_tile * tile) xe_vm_pdp4_descriptor() argument
1752 to_wait_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) to_wait_exec_queue() argument
1783 struct xe_vm *vm; xe_vm_create_ioctl() local
1868 struct xe_vm *vm; xe_vm_destroy_ioctl() local
1897 prep_vma_destroy(struct xe_vm * vm,struct xe_vma * vma,bool post_commit) prep_vma_destroy() argument
1962 vm_bind_ioctl_ops_create(struct xe_vm * vm,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 addr,u64 range,u32 operation,u32 flags,u32 prefetch_region,u16 pat_index) vm_bind_ioctl_ops_create() argument
2038 new_vma(struct xe_vm * vm,struct drm_gpuva_op_map * op,u16 pat_index,unsigned int flags) new_vma() argument
2122 xe_vma_op_commit(struct xe_vm * vm,struct xe_vma_op * op) xe_vma_op_commit() argument
2185 vm_bind_ioctl_ops_parse(struct xe_vm * vm,struct drm_gpuva_ops * ops,struct xe_vma_ops * vops) vm_bind_ioctl_ops_parse() argument
2342 xe_vma_op_unwind(struct xe_vm * vm,struct xe_vma_op * op,bool post_commit,bool prev_post_commit,bool next_post_commit) xe_vma_op_unwind() argument
2397 vm_bind_ioctl_ops_unwind(struct xe_vm * vm,struct drm_gpuva_ops ** ops,int num_ops_list) vm_bind_ioctl_ops_unwind() argument
2425 struct xe_vm *vm = xe_vma_vm(vma); vma_lock_and_validate() local
2454 op_lock_and_prep(struct drm_exec * exec,struct xe_vm * vm,struct xe_vma_op * op) op_lock_and_prep() argument
2510 vm_bind_ioctl_ops_lock_and_prep(struct drm_exec * exec,struct xe_vm * vm,struct xe_vma_ops * vops) vm_bind_ioctl_ops_lock_and_prep() argument
2567 vm_ops_setup_tile_args(struct xe_vm * vm,struct xe_vma_ops * vops) vm_ops_setup_tile_args() argument
2593 ops_execute(struct xe_vm * vm,struct xe_vma_ops * vops) ops_execute() argument
2687 op_add_ufence(struct xe_vm * vm,struct xe_vma_op * op,struct xe_user_fence * ufence) op_add_ufence() argument
2710 vm_bind_ioctl_ops_fini(struct xe_vm * vm,struct xe_vma_ops * vops,struct dma_fence * fence) vm_bind_ioctl_ops_fini() argument
2737 vm_bind_ioctl_ops_execute(struct xe_vm * vm,struct xe_vma_ops * vops) vm_bind_ioctl_ops_execute() argument
2899 vm_bind_ioctl_signal_fences(struct xe_vm * vm,struct xe_exec_queue * q,struct xe_sync_entry * syncs,int num_syncs) vm_bind_ioctl_signal_fences() argument
2922 xe_vma_ops_init(struct xe_vma_ops * vops,struct xe_vm * vm,struct xe_exec_queue * q,struct xe_sync_entry * syncs,u32 num_syncs) xe_vma_ops_init() argument
2991 struct xe_vm *vm; xe_vm_bind_ioctl() local
3204 xe_vm_lock(struct xe_vm * vm,bool intr) xe_vm_lock() argument
3218 xe_vm_unlock(struct xe_vm * vm) xe_vm_unlock() argument
3312 xe_vm_snapshot_capture(struct xe_vm * vm) xe_vm_snapshot_capture() argument
[all...]
H A Dxe_exec_queue.c51 __xe_exec_queue_alloc(struct xe_device * xe,struct xe_vm * vm,u32 logical_mask,u16 width,struct xe_hw_engine * hwe,u32 flags,u64 extensions) __xe_exec_queue_alloc() argument
113 struct xe_vm *vm = q->vm; __xe_exec_queue_init() local
148 xe_exec_queue_create(struct xe_device * xe,struct xe_vm * vm,u32 logical_mask,u16 width,struct xe_hw_engine * hwe,u32 flags,u64 extensions) xe_exec_queue_create() argument
173 xe_exec_queue_create_class(struct xe_device * xe,struct xe_gt * gt,struct xe_vm * vm,enum xe_engine_class class,u32 flags,u64 extensions) xe_exec_queue_create_class() argument
547 struct xe_vm *vm; xe_exec_queue_create_ioctl() local
858 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_lockdep_assert() argument
873 xe_exec_queue_last_fence_put(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_put() argument
904 xe_exec_queue_last_fence_get(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get() argument
931 xe_exec_queue_last_fence_get_for_resume(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get_for_resume() argument
955 xe_exec_queue_last_fence_set(struct xe_exec_queue * q,struct xe_vm * vm,struct dma_fence * fence) xe_exec_queue_last_fence_set() argument
972 xe_exec_queue_last_fence_test_dep(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_test_dep() argument
[all...]
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_vm.c122 struct amdgpu_vm *vm; member
141 int amdgpu_vm_set_pasid(struct amdgpu_device *adev, struct amdgpu_vm *vm, in amdgpu_vm_set_pasid() argument
180 struct amdgpu_vm *vm = vm_bo->vm; amdgpu_vm_bo_evicted() local
294 amdgpu_vm_bo_reset_state_machine(struct amdgpu_vm * vm) amdgpu_vm_bo_reset_state_machine() argument
324 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_shared() local
373 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_stats_locked() local
412 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_stats() local
430 amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base * base,struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_base_init() argument
479 amdgpu_vm_lock_pd(struct amdgpu_vm * vm,struct drm_exec * exec,unsigned int num_fences) amdgpu_vm_lock_pd() argument
497 amdgpu_vm_move_to_lru_tail(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_move_to_lru_tail() argument
506 amdgpu_vm_init_entities(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_init_entities() argument
526 amdgpu_vm_fini_entities(struct amdgpu_vm * vm) amdgpu_vm_fini_entities() argument
541 amdgpu_vm_generation(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_generation() argument
572 amdgpu_vm_validate(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket,int (* validate)(void * p,struct amdgpu_bo * bo),void * param) amdgpu_vm_validate() argument
659 amdgpu_vm_ready(struct amdgpu_vm * vm) amdgpu_vm_ready() argument
875 amdgpu_vm_bo_find(struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_find() argument
929 amdgpu_vm_update_pdes(struct amdgpu_device * adev,struct amdgpu_vm * vm,bool immediate) amdgpu_vm_update_pdes() argument
1014 struct amdgpu_vm *vm = params->vm; amdgpu_vm_tlb_flush() local
1064 amdgpu_vm_update_range(struct amdgpu_device * adev,struct amdgpu_vm * vm,bool immediate,bool unlocked,bool flush_tlb,bool allow_override,struct amdgpu_sync * sync,uint64_t start,uint64_t last,uint64_t flags,uint64_t offset,uint64_t vram_base,struct ttm_resource * res,dma_addr_t * pages_addr,struct dma_fence ** fence) amdgpu_vm_update_range() argument
1197 amdgpu_vm_get_memory(struct amdgpu_vm * vm,struct amdgpu_mem_stats stats[__AMDGPU_PL_NUM]) amdgpu_vm_get_memory() argument
1221 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_update() local
1457 amdgpu_vm_free_mapping(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_bo_va_mapping * mapping,struct dma_fence * fence) amdgpu_vm_free_mapping() argument
1474 amdgpu_vm_prt_fini(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_prt_fini() argument
1503 amdgpu_vm_clear_freed(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct dma_fence ** fence) amdgpu_vm_clear_freed() argument
1565 amdgpu_vm_handle_moved(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket) amdgpu_vm_handle_moved() argument
1643 amdgpu_vm_flush_compute_tlb(struct amdgpu_device * adev,struct amdgpu_vm * vm,uint32_t flush_type,uint32_t xcc_mask) amdgpu_vm_flush_compute_tlb() argument
1690 amdgpu_vm_bo_add(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_add() argument
1733 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_insert_map() local
1804 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_map() local
1913 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_unmap() local
1963 amdgpu_vm_bo_clear_mappings(struct amdgpu_device * adev,struct amdgpu_vm * vm,uint64_t saddr,uint64_t size) amdgpu_vm_bo_clear_mappings() argument
2081 amdgpu_vm_bo_lookup_mapping(struct amdgpu_vm * vm,uint64_t addr) amdgpu_vm_bo_lookup_mapping() argument
2095 amdgpu_vm_bo_trace_cs(struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket) amdgpu_vm_bo_trace_cs() argument
2132 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_del() local
2226 struct amdgpu_vm *vm = bo_base->vm; amdgpu_vm_bo_invalidate() local
2261 struct amdgpu_vm *vm = bo_base->vm; amdgpu_vm_bo_move() local
2393 amdgpu_vm_wait_idle(struct amdgpu_vm * vm,long timeout) amdgpu_vm_wait_idle() argument
2414 struct amdgpu_vm *vm; amdgpu_vm_get_vm_from_pasid() local
2445 amdgpu_vm_get_task_info_vm(struct amdgpu_vm * vm) amdgpu_vm_get_task_info_vm() argument
2473 amdgpu_vm_create_task_info(struct amdgpu_vm * vm) amdgpu_vm_create_task_info() argument
2488 amdgpu_vm_set_task_info(struct amdgpu_vm * vm) amdgpu_vm_set_task_info() argument
2518 amdgpu_vm_init(struct amdgpu_device * adev,struct amdgpu_vm * vm,int32_t xcp_id) amdgpu_vm_init() argument
2635 amdgpu_vm_make_compute(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_make_compute() argument
2685 amdgpu_vm_release_compute(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_release_compute() argument
2691 amdgpu_vm_stats_is_zero(struct amdgpu_vm * vm) amdgpu_vm_stats_is_zero() argument
2710 amdgpu_vm_fini(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_fini() argument
2913 struct amdgpu_vm *vm; amdgpu_vm_handle_fault() local
3003 amdgpu_debugfs_vm_bo_info(struct amdgpu_vm * vm,struct seq_file * m) amdgpu_debugfs_vm_bo_info() argument
3106 struct amdgpu_vm *vm; amdgpu_vm_update_fault_cache() local
3156 amdgpu_vm_is_bo_always_valid(struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_is_bo_always_valid() argument
[all...]
H A Damdgpu_vm_pt.c157 struct amdgpu_vm *vm, uint64_t start, in amdgpu_vm_pt_start() argument
286 struct amdgpu_vm *vm, in amdgpu_vm_pt_first_dfs() argument
340 #define for_each_amdgpu_vm_pt_dfs_safe(adev, vm, start, cursor, entry) \ argument
359 int amdgpu_vm_pt_clear(struct amdgpu_device *adev, struct amdgpu_vm *vm, in amdgpu_vm_pt_clear() argument
438 amdgpu_vm_pt_create(struct amdgpu_device * adev,struct amdgpu_vm * vm,int level,bool immediate,struct amdgpu_bo_vm ** vmbo,int32_t xcp_id) amdgpu_vm_pt_create() argument
494 amdgpu_vm_pt_alloc(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_vm_pt_cursor * cursor,bool immediate) amdgpu_vm_pt_alloc() argument
553 struct amdgpu_vm *vm; amdgpu_vm_pt_free_work() local
583 struct amdgpu_vm *vm = params->vm; amdgpu_vm_pt_free_list() local
633 amdgpu_vm_pt_free_root(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_pt_free_root() argument
657 struct amdgpu_vm *vm = params->vm; amdgpu_vm_pde_update() local
908 struct amdgpu_vm *vm = params->vm; amdgpu_vm_ptes_update() local
972 amdgpu_vm_pt_map_tables(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_pt_map_tables() argument
[all...]
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_gtt.c41 struct drm_i915_gem_object *alloc_pt_lmem(struct i915_address_space *vm, int sz) in alloc_pt_lmem()
75 struct drm_i915_gem_object *alloc_pt_dma(struct i915_address_space *vm, int sz) in alloc_pt_dma()
99 int map_pt_dma(struct i915_address_space *vm, struct drm_i915_gem_object *obj) in map_pt_dma()
123 int map_pt_dma_locked(struct i915_address_space *vm, struct drm_i915_gem_object *obj) in map_pt_dma_locked()
183 static void __i915_vm_close(struct i915_address_space *vm) in __i915_vm_close()
198 int i915_vm_lock_objects(struct i915_address_space *vm, in i915_vm_lock_objects()
211 void i915_address_space_fini(struct i915_address_space *vm) in i915_address_space_fini()
226 struct i915_address_space *vm = in i915_vm_resv_release() local
237 struct i915_address_space *vm = in __i915_vm_release() local
253 struct i915_address_space *vm = in i915_vm_release() local
[all …]
H A Dgen8_ppgtt.c180 static unsigned int gen8_pd_top_count(const struct i915_address_space *vm) in gen8_pd_top_count()
188 gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) in gen8_pdp_for_page_index()
199 gen8_pdp_for_page_address(struct i915_address_space * const vm, const u64 addr) in gen8_pdp_for_page_address()
204 static void __gen8_ppgtt_cleanup(struct i915_address_space *vm, in __gen8_ppgtt_cleanup()
222 static void gen8_ppgtt_cleanup(struct i915_address_space *vm) in gen8_ppgtt_cleanup()
239 static u64 __gen8_ppgtt_clear(struct i915_address_space * const vm, in __gen8_ppgtt_clear()
307 static void gen8_ppgtt_clear(struct i915_address_space *vm, in gen8_ppgtt_clear()
322 static void __gen8_ppgtt_alloc(struct i915_address_space * const vm, in __gen8_ppgtt_alloc()
390 static void gen8_ppgtt_alloc(struct i915_address_space *vm, in gen8_ppgtt_alloc()
406 static void __gen8_ppgtt_foreach(struct i915_address_space *vm, in __gen8_ppgtt_foreach()
[all …]
H A Dintel_ggtt.c115 void i915_ggtt_suspend_vm(struct i915_address_space *vm, bool evict_all) in i915_ggtt_suspend_vm()
438 static void gen8_ggtt_insert_page(struct i915_address_space *vm, in gen8_ggtt_insert_page()
453 static void gen8_ggtt_insert_page_bind(struct i915_address_space *vm, in gen8_ggtt_insert_page_bind()
468 static void gen8_ggtt_insert_entries(struct i915_address_space *vm, in gen8_ggtt_insert_entries()
507 static bool __gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, in __gen8_ggtt_insert_entries_bind()
538 static void gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, in gen8_ggtt_insert_entries_bind()
551 static void gen8_ggtt_clear_range(struct i915_address_space *vm, in gen8_ggtt_clear_range()
572 static void gen8_ggtt_scratch_range_bind(struct i915_address_space *vm, in gen8_ggtt_scratch_range_bind()
593 static void gen6_ggtt_insert_page(struct i915_address_space *vm, in gen6_ggtt_insert_page()
614 static void gen6_ggtt_insert_entries(struct i915_address_space *vm, in gen6_ggtt_insert_entries()
[all …]
H A Dintel_ppgtt.c16 struct i915_page_table *alloc_pt(struct i915_address_space *vm, int sz) in alloc_pt()
53 struct i915_page_directory *alloc_pd(struct i915_address_space *vm) in alloc_pd()
71 void free_px(struct i915_address_space *vm, struct i915_page_table *pt, int lvl) in free_px()
182 void ppgtt_bind_vma(struct i915_address_space *vm, in ppgtt_bind_vma()
207 void ppgtt_unbind_vma(struct i915_address_space *vm, in ppgtt_unbind_vma()
223 int i915_vm_alloc_pt_stash(struct i915_address_space *vm, in i915_vm_alloc_pt_stash()
276 int i915_vm_map_pt_stash(struct i915_address_space *vm, in i915_vm_map_pt_stash()
293 void i915_vm_free_pt_stash(struct i915_address_space *vm, in i915_vm_free_pt_stash()
H A Dintel_gtt.h367 struct i915_address_space vm; member
410 struct i915_address_space vm; member
415 #define i915_is_ggtt(vm) ((vm)->is_ggtt) argument
416 #define i915_is_dpt(vm) ((vm)->is_dpt) argument
417 #define i915_is_ggtt_or_dpt(vm) (i915_is_ggtt(vm) || i915_is_dpt(vm)) argument
425 i915_vm_is_4lvl(const struct i915_address_space *vm) in i915_vm_is_4lvl()
431 i915_vm_has_scratch_64K(struct i915_address_space *vm) in i915_vm_has_scratch_64K()
436 static inline u64 i915_vm_min_alignment(struct i915_address_space *vm, in i915_vm_min_alignment()
446 static inline u64 i915_vm_obj_min_alignment(struct i915_address_space *vm, in i915_vm_obj_min_alignment()
456 i915_vm_has_cache_coloring(struct i915_address_space *vm) in i915_vm_has_cache_coloring()
[all …]
H A Dintel_ggtt_gmch.c18 static void gmch_ggtt_insert_page(struct i915_address_space *vm, in gmch_ggtt_insert_page()
30 static void gmch_ggtt_insert_entries(struct i915_address_space *vm, in gmch_ggtt_insert_entries()
47 static void gmch_ggtt_clear_range(struct i915_address_space *vm, in gmch_ggtt_clear_range()
53 static void gmch_ggtt_remove(struct i915_address_space *vm) in gmch_ggtt_remove()
/linux/drivers/gpu/drm/panthor/
H A Dpanthor_mmu.c46 struct panthor_vm *vm; member
405 /** @vm member
101 } vm; global() member
437 struct panthor_vm *vm = cookie; alloc_pt() local
489 struct panthor_vm *vm = cookie; free_pt() local
595 mmu_hw_do_operation(struct panthor_vm * vm,u64 iova,u64 size,u32 op) mmu_hw_do_operation() argument
666 panthor_vm_has_unhandled_faults(struct panthor_vm * vm) panthor_vm_has_unhandled_faults() argument
677 panthor_vm_is_unusable(struct panthor_vm * vm) panthor_vm_is_unusable() argument
682 panthor_vm_release_as_locked(struct panthor_vm * vm) panthor_vm_release_as_locked() argument
706 panthor_vm_active(struct panthor_vm * vm) panthor_vm_active() argument
815 panthor_vm_idle(struct panthor_vm * vm) panthor_vm_idle() argument
829 panthor_vm_page_size(struct panthor_vm * vm) panthor_vm_page_size() argument
837 panthor_vm_stop(struct panthor_vm * vm) panthor_vm_stop() argument
842 panthor_vm_start(struct panthor_vm * vm) panthor_vm_start() argument
853 panthor_vm_as(struct panthor_vm * vm) panthor_vm_as() argument
879 panthor_vm_flush_range(struct panthor_vm * vm,u64 iova,u64 size) panthor_vm_flush_range() argument
903 panthor_vm_flush_all(struct panthor_vm * vm) panthor_vm_flush_all() argument
908 panthor_vm_unmap_pages(struct panthor_vm * vm,u64 iova,u64 size) panthor_vm_unmap_pages() argument
937 panthor_vm_map_pages(struct panthor_vm * vm,u64 iova,int prot,struct sg_table * sgt,u64 offset,u64 size) panthor_vm_map_pages() argument
1035 panthor_vm_alloc_va(struct panthor_vm * vm,u64 va,u64 size,struct drm_mm_node * va_node) panthor_vm_alloc_va() argument
1069 panthor_vm_free_va(struct panthor_vm * vm,struct drm_mm_node * va_node) panthor_vm_free_va() argument
1079 struct drm_gpuvm *vm = vm_bo->vm; panthor_vm_bo_put() local
1113 panthor_vm_cleanup_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm) panthor_vm_cleanup_op_ctx() argument
1199 panthor_vm_prepare_map_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm,struct panthor_gem_object * bo,u64 offset,u64 size,u64 va,u32 flags) panthor_vm_prepare_map_op_ctx() argument
1327 panthor_vm_prepare_unmap_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm,u64 va,u64 size) panthor_vm_prepare_unmap_op_ctx() argument
1379 panthor_vm_prepare_sync_only_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm) panthor_vm_prepare_sync_only_op_ctx() argument
1401 panthor_vm_get_bo_for_va(struct panthor_vm * vm,u64 va,u64 * bo_offset) panthor_vm_get_bo_for_va() argument
1506 struct panthor_vm *vm; panthor_vm_pool_create_vm() local
1531 panthor_vm_destroy(struct panthor_vm * vm) panthor_vm_destroy() argument
1564 struct panthor_vm *vm; panthor_vm_pool_destroy_vm() local
1583 struct panthor_vm *vm; panthor_vm_pool_get_vm() local
1603 struct panthor_vm *vm; panthor_vm_pool_destroy() local
1745 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_suspend() local
1788 struct panthor_vm *vm; panthor_mmu_pre_reset() local
1808 struct panthor_vm *vm; panthor_mmu_post_reset() local
1819 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_post_reset() local
1840 struct panthor_vm *vm = container_of(gpuvm, struct panthor_vm, base); panthor_vm_free() local
1888 panthor_vm_put(struct panthor_vm * vm) panthor_vm_put() argument
1899 panthor_vm_get(struct panthor_vm * vm) panthor_vm_get() argument
1921 panthor_vm_get_heap_pool(struct panthor_vm * vm,bool create) panthor_vm_get_heap_pool() argument
1986 panthor_vma_link(struct panthor_vm * vm,struct panthor_vma * vma,struct drm_gpuvm_bo * vm_bo) panthor_vma_link() argument
1998 panthor_vma_unlink(struct panthor_vm * vm,struct panthor_vma * vma) panthor_vma_unlink() argument
2029 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_map() local
2058 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_remap() local
2106 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_unmap() local
2132 panthor_vm_resv(struct panthor_vm * vm) panthor_vm_resv() argument
2137 panthor_vm_root_gem(struct panthor_vm * vm) panthor_vm_root_gem() argument
2146 panthor_vm_exec_op(struct panthor_vm * vm,struct panthor_vm_op_ctx * op,bool flag_vm_unusable_on_failure) panthor_vm_exec_op() argument
2280 struct panthor_vm *vm; panthor_vm_create() local
2385 panthor_vm_bind_prepare_op_ctx(struct drm_file * file,struct panthor_vm * vm,const struct drm_panthor_vm_bind_op * op,struct panthor_vm_op_ctx * op_ctx) panthor_vm_bind_prepare_op_ctx() argument
2457 panthor_vm_bind_job_create(struct drm_file * file,struct panthor_vm * vm,const struct drm_panthor_vm_bind_op * op) panthor_vm_bind_job_create() argument
2543 panthor_vm_update_resvs(struct panthor_vm * vm,struct drm_exec * exec,struct dma_fence * fence,enum dma_resv_usage private_usage,enum dma_resv_usage extobj_usage) panthor_vm_update_resvs() argument
2560 panthor_vm_bind_exec_sync_op(struct drm_file * file,struct panthor_vm * vm,struct drm_panthor_vm_bind_op * op) panthor_vm_bind_exec_sync_op() argument
2598 panthor_vm_map_bo_range(struct panthor_vm * vm,struct panthor_gem_object * bo,u64 offset,u64 size,u64 va,u32 flags) panthor_vm_map_bo_range() argument
2625 panthor_vm_unmap_range(struct panthor_vm * vm,u64 va,u64 size) panthor_vm_unmap_range() argument
2653 panthor_vm_prepare_mapped_bos_resvs(struct drm_exec * exec,struct panthor_vm * vm,u32 slot_count) panthor_vm_prepare_mapped_bos_resvs() argument
2680 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_unplug() local
2750 show_vm_gpuvas(struct panthor_vm * vm,struct seq_file * m) show_vm_gpuvas() argument
2767 struct panthor_vm *vm; show_each_vm() local
[all...]
/linux/drivers/gpu/drm/i915/display/
H A Dintel_dpt.c17 struct i915_address_space vm; member
24 #define i915_is_dpt(vm) ((vm)->is_dpt) argument
27 i915_vm_to_dpt(struct i915_address_space *vm) in i915_vm_to_dpt() argument
41 dpt_insert_page(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) dpt_insert_page() argument
54 dpt_insert_entries(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) dpt_insert_entries() argument
76 dpt_clear_range(struct i915_address_space * vm,u64 start,u64 length) dpt_clear_range() argument
81 dpt_bind_vma(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) dpt_bind_vma() argument
111 dpt_unbind_vma(struct i915_address_space * vm,struct i915_vma_resource * vma_res) dpt_unbind_vma() argument
117 dpt_cleanup(struct i915_address_space * vm) dpt_cleanup() argument
124 intel_dpt_pin_to_ggtt(struct i915_address_space * vm,unsigned int alignment) intel_dpt_pin_to_ggtt() argument
176 intel_dpt_unpin_from_ggtt(struct i915_address_space * vm) intel_dpt_unpin_from_ggtt() argument
248 struct i915_address_space *vm; intel_dpt_create() local
312 intel_dpt_destroy(struct i915_address_space * vm) intel_dpt_destroy() argument
[all...]
/linux/drivers/gpu/drm/i915/selftests/
H A Dmock_gtt.c27 static void mock_insert_page(struct i915_address_space *vm, in mock_insert_page()
35 static void mock_insert_entries(struct i915_address_space *vm, in mock_insert_entries()
41 static void mock_bind_ppgtt(struct i915_address_space *vm, in mock_bind_ppgtt()
51 static void mock_unbind_ppgtt(struct i915_address_space *vm, in mock_unbind_ppgtt()
56 static void mock_cleanup(struct i915_address_space *vm) in mock_cleanup()
60 static void mock_clear_range(struct i915_address_space *vm, in mock_clear_range()
94 static void mock_bind_ggtt(struct i915_address_space *vm, in mock_bind_ggtt()
102 static void mock_unbind_ggtt(struct i915_address_space *vm, in mock_unbind_ggtt()
H A Di915_gem_gtt.c244 lowlevel_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) lowlevel_hole() argument
391 close_object_list(struct list_head * objects,struct i915_address_space * vm) close_object_list() argument
408 fill_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) fill_hole() argument
633 walk_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) walk_hole() argument
716 pot_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) pot_hole() argument
791 drunk_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) drunk_hole() argument
901 __shrink_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) __shrink_hole() argument
979 shrink_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) shrink_hole() argument
1001 shrink_boom(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) shrink_boom() argument
1079 misaligned_case(struct i915_address_space * vm,struct intel_memory_region * mr,u64 addr,u64 size,unsigned long flags) misaligned_case() argument
1144 misaligned_pin(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) misaligned_pin() argument
1191 exercise_ppgtt(struct drm_i915_private * dev_priv,int (* func)(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time)) exercise_ppgtt() argument
1277 exercise_ggtt(struct drm_i915_private * i915,int (* func)(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time)) exercise_ggtt() argument
1454 exercise_mock(struct drm_i915_private * i915,int (* func)(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time)) exercise_mock() argument
1459 struct i915_address_space *vm; exercise_mock() local
1506 struct i915_address_space *vm = vma->vm; reserve_gtt_with_resource() local
1687 struct i915_address_space *vm = vma->vm; insert_gtt_with_resource() local
[all...]
/linux/drivers/virt/acrn/
H A Dirqfd.c32 struct acrn_vm *vm; member
43 struct acrn_vm *vm = irqfd->vm; in acrn_irqfd_inject() local
65 struct acrn_vm *vm; in hsm_irqfd_shutdown_work() local
81 struct acrn_vm *vm; in hsm_irqfd_wakeup() local
110 static int acrn_irqfd_assign(struct acrn_vm *vm, struct acrn_irqfd *args) in acrn_irqfd_assign()
172 static int acrn_irqfd_deassign(struct acrn_vm *vm, in acrn_irqfd_deassign()
195 int acrn_irqfd_config(struct acrn_vm *vm, struct acrn_irqfd *args) in acrn_irqfd_config()
207 int acrn_irqfd_init(struct acrn_vm *vm) in acrn_irqfd_init()
219 void acrn_irqfd_deinit(struct acrn_vm *vm) in acrn_irqfd_deinit()
H A Dioeventfd.c43 static void acrn_ioeventfd_shutdown(struct acrn_vm *vm, struct hsm_ioeventfd *p) in acrn_ioeventfd_shutdown()
52 static bool hsm_ioeventfd_is_conflict(struct acrn_vm *vm, in hsm_ioeventfd_is_conflict()
76 static int acrn_ioeventfd_assign(struct acrn_vm *vm, in acrn_ioeventfd_assign()
147 static int acrn_ioeventfd_deassign(struct acrn_vm *vm, in acrn_ioeventfd_deassign()
173 static struct hsm_ioeventfd *hsm_ioeventfd_match(struct acrn_vm *vm, u64 addr, in hsm_ioeventfd_match()
232 int acrn_ioeventfd_config(struct acrn_vm *vm, struct acrn_ioeventfd *args) in acrn_ioeventfd_config()
244 int acrn_ioeventfd_init(struct acrn_vm *vm) in acrn_ioeventfd_init()
263 void acrn_ioeventfd_deinit(struct acrn_vm *vm) in acrn_ioeventfd_deinit()
H A Dioreq.c39 static int ioreq_complete_request(struct acrn_vm *vm, u16 vcpu, in ioreq_complete_request()
93 int acrn_ioreq_request_default_complete(struct acrn_vm *vm, u16 vcpu) in acrn_ioreq_request_default_complete()
212 void acrn_ioreq_request_clear(struct acrn_vm *vm) in acrn_ioreq_request_clear()
312 static bool handle_cf8cfc(struct acrn_vm *vm, in handle_cf8cfc()
381 static struct acrn_ioreq_client *find_ioreq_client(struct acrn_vm *vm, in find_ioreq_client()
415 struct acrn_ioreq_client *acrn_ioreq_client_create(struct acrn_vm *vm, in acrn_ioreq_client_create()
468 struct acrn_vm *vm = client->vm; in acrn_ioreq_client_destroy() local
497 static int acrn_ioreq_dispatch(struct acrn_vm *vm) in acrn_ioreq_dispatch()
546 struct acrn_vm *vm; in ioreq_dispatcher() local
596 int acrn_ioreq_init(struct acrn_vm *vm, u64 buf_vma) in acrn_ioreq_init()
[all …]
H A Dmm.c19 static int modify_region(struct acrn_vm *vm, struct vm_memory_region_op *region) in modify_region()
52 int acrn_mm_region_add(struct acrn_vm *vm, u64 user_gpa, u64 service_gpa, in acrn_mm_region_add()
85 int acrn_mm_region_del(struct acrn_vm *vm, u64 user_gpa, u64 size) in acrn_mm_region_del()
108 int acrn_vm_memseg_map(struct acrn_vm *vm, struct acrn_vm_memmap *memmap) in acrn_vm_memseg_map()
131 int acrn_vm_memseg_unmap(struct acrn_vm *vm, struct acrn_vm_memmap *memmap) in acrn_vm_memseg_unmap()
156 int acrn_vm_ram_map(struct acrn_vm *vm, struct acrn_vm_memmap *memmap) in acrn_vm_ram_map()
354 void acrn_vm_all_ram_unmap(struct acrn_vm *vm) in acrn_vm_all_ram_unmap()
/linux/drivers/gpu/drm/lima/
H A Dlima_vm.c18 struct lima_vm *vm; member
35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) in lima_vm_unmap_range()
47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page()
77 lima_vm_bo_find(struct lima_vm *vm, struct lima_bo *bo) in lima_vm_bo_find()
91 int lima_vm_bo_add(struct lima_vm *vm, struct lima_bo *bo, bool create) in lima_vm_bo_add()
155 void lima_vm_bo_del(struct lima_vm *vm, struct lima_bo *bo) in lima_vm_bo_del()
185 u32 lima_vm_get_va(struct lima_vm *vm, struct lima_bo *bo) in lima_vm_get_va()
202 struct lima_vm *vm; in lima_vm_create() local
237 struct lima_vm *vm = container_of(kref, struct lima_vm, refcount); in lima_vm_release() local
254 void lima_vm_print(struct lima_vm *vm) in lima_vm_print()
[all …]
/linux/sound/pci/ctxfi/
H A Dctvmem.c30 get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc) in get_vm_block()
75 static void put_vm_block(struct ct_vm *vm, struct ct_vm_block *block) in put_vm_block()
124 ct_vm_map(struct ct_vm *vm, struct snd_pcm_substream *substream, int size) in ct_vm_map()
152 static void ct_vm_unmap(struct ct_vm *vm, struct ct_vm_block *block) in ct_vm_unmap()
164 ct_get_ptp_phys(struct ct_vm *vm, int index) in ct_get_ptp_phys()
171 struct ct_vm *vm; in ct_vm_create() local
215 void ct_vm_destroy(struct ct_vm *vm) in ct_vm_destroy()
/linux/tools/testing/selftests/kvm/lib/
H A Dmemstress.c88 void memstress_setup_vcpus(struct kvm_vm *vm, int nr_vcpus, in memstress_setup_vcpus()
130 struct kvm_vm *vm; in memstress_create_vm() local
234 void memstress_destroy_vm(struct kvm_vm *vm) in memstress_destroy_vm()
239 void memstress_set_write_percent(struct kvm_vm *vm, uint32_t write_percent) in memstress_set_write_percent()
245 void memstress_set_random_access(struct kvm_vm *vm, bool random_access) in memstress_set_random_access()
256 void __weak memstress_setup_nested(struct kvm_vm *vm, int nr_vcpus, struct kvm_vcpu **vcpus) in memstress_setup_nested()
322 static void toggle_dirty_logging(struct kvm_vm *vm, int slots, bool enable) in toggle_dirty_logging()
334 void memstress_enable_dirty_logging(struct kvm_vm *vm, int slots) in memstress_enable_dirty_logging()
339 void memstress_disable_dirty_logging(struct kvm_vm *vm, int slots) in memstress_disable_dirty_logging()
344 void memstress_get_dirty_log(struct kvm_vm *vm, unsigned long *bitmaps[], int slots) in memstress_get_dirty_log()
[all …]
/linux/drivers/gpu/drm/radeon/
H A Dradeon_vm.c130 struct radeon_vm *vm, in radeon_vm_get_bos()
177 struct radeon_vm *vm, int ring) in radeon_vm_grab_id()
236 struct radeon_vm *vm, in radeon_vm_flush()
268 struct radeon_vm *vm, in radeon_vm_fence()
292 struct radeon_bo_va *radeon_vm_bo_find(struct radeon_vm *vm, in radeon_vm_bo_find()
319 struct radeon_vm *vm, in radeon_vm_bo_add()
451 struct radeon_vm *vm = bo_va->vm; in radeon_vm_bo_set_addr() local
639 struct radeon_vm *vm) in radeon_vm_update_page_directory()
813 struct radeon_vm *vm, in radeon_vm_update_ptes()
884 static void radeon_vm_fence_pts(struct radeon_vm *vm, in radeon_vm_fence_pts()
[all …]
/linux/drivers/gpu/drm/i915/
H A Di915_vma_resource.c105 struct i915_address_space *vm; in __i915_vma_resource_unhold() local
177 struct i915_address_space *vm = vma_res->vm; in i915_vma_resource_unbind_work() local
233 struct i915_address_space *vm = vma_res->vm; in i915_vma_resource_unbind() local
272 i915_vma_resource_color_adjust_range(struct i915_address_space *vm, in i915_vma_resource_color_adjust_range()
295 int i915_vma_resource_bind_dep_sync(struct i915_address_space *vm, in i915_vma_resource_bind_dep_sync()
330 void i915_vma_resource_bind_dep_sync_all(struct i915_address_space *vm) in i915_vma_resource_bind_dep_sync_all()
379 int i915_vma_resource_bind_dep_await(struct i915_address_space *vm, in i915_vma_resource_bind_dep_await()

123456789