/linux/tools/testing/selftests/kvm/s390/ |
H A D | cmma_test.c | 97 static void create_main_memslot(struct kvm_vm *vm) in create_main_memslot() 107 static void create_test_memslot(struct kvm_vm *vm) in create_test_memslot() 119 static void create_memslots(struct kvm_vm *vm) in create_memslots() 139 static void finish_vm_setup(struct kvm_vm *vm) in finish_vm_setup() 153 struct kvm_vm *vm; in create_vm_two_memslots() local 164 static void enable_cmma(struct kvm_vm *vm) in enable_cmma() 172 static void enable_dirty_tracking(struct kvm_vm *vm) in enable_dirty_tracking() 178 static int __enable_migration_mode(struct kvm_vm *vm) in __enable_migration_mode() 187 static void enable_migration_mode(struct kvm_vm *vm) in enable_migration_mode() 194 static bool is_migration_mode_on(struct kvm_vm *vm) in is_migration_mode_on() [all …]
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_vm.h | 37 return vm; in xe_vm_put() argument 52 return !vm argument 31 xe_vm_get(struct xe_vm * vm) xe_vm_get() argument 46 xe_vm_is_closed(struct xe_vm * vm) xe_vm_is_closed() argument 57 xe_vm_is_closed_or_banned(struct xe_vm * vm) xe_vm_is_closed_or_banned() argument 72 xe_vm_has_scratch(const struct xe_vm * vm) xe_vm_has_scratch() argument 186 xe_vm_in_fault_mode(struct xe_vm * vm) xe_vm_in_fault_mode() argument 191 xe_vm_in_lr_mode(struct xe_vm * vm) xe_vm_in_lr_mode() argument 196 xe_vm_in_preempt_fence_mode(struct xe_vm * vm) xe_vm_in_preempt_fence_mode() argument 216 xe_vm_queue_rebind_worker(struct xe_vm * vm) xe_vm_queue_rebind_worker() argument 231 xe_vm_reactivate_rebind(struct xe_vm * vm) xe_vm_reactivate_rebind() argument 256 xe_vm_resv(struct xe_vm * vm) xe_vm_resv() argument 267 xe_vm_assert_held(vm) global() argument [all...] |
H A D | xe_vm.c | 42 xe_vm_obj(struct xe_vm * vm) xe_vm_obj() argument 69 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_userptr_pin_pages() local 78 preempt_fences_waiting(struct xe_vm * vm) preempt_fences_waiting() argument 104 alloc_preempt_fences(struct xe_vm * vm,struct list_head * list,unsigned int * count) alloc_preempt_fences() argument 125 wait_for_existing_preempt_fences(struct xe_vm * vm) wait_for_existing_preempt_fences() argument 147 xe_vm_is_idle(struct xe_vm * vm) xe_vm_is_idle() argument 160 arm_preempt_fences(struct xe_vm * vm,struct list_head * list) arm_preempt_fences() argument 179 add_preempt_fences(struct xe_vm * vm,struct xe_bo * bo) add_preempt_fences() argument 203 resume_and_reinstall_preempt_fences(struct xe_vm * vm,struct drm_exec * exec) resume_and_reinstall_preempt_fences() argument 219 xe_vm_add_compute_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) xe_vm_add_compute_exec_queue() argument 280 xe_vm_remove_compute_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) xe_vm_remove_compute_exec_queue() argument 309 __xe_vm_userptr_needs_repin(struct xe_vm * vm) __xe_vm_userptr_needs_repin() argument 327 xe_vm_kill(struct xe_vm * vm,bool unlocked) xe_vm_kill() argument 383 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); xe_gpuvm_validate() local 415 xe_vm_validate_rebind(struct xe_vm * vm,struct drm_exec * exec,unsigned int num_fences) xe_vm_validate_rebind() argument 441 xe_preempt_work_begin(struct drm_exec * exec,struct xe_vm * vm,bool * done) xe_preempt_work_begin() argument 480 struct xe_vm *vm = container_of(w, struct xe_vm, preempt.rebind_work); preempt_rebind_work_func() local 582 __vma_userptr_invalidate(struct xe_vm * vm,struct xe_userptr_vma * uvma) __vma_userptr_invalidate() argument 633 struct xe_vm *vm = xe_vma_vm(vma); vma_userptr_invalidate() local 668 struct xe_vm *vm = xe_vma_vm(&uvma->vma); xe_vma_userptr_force_invalidate() local 687 xe_vm_userptr_pin(struct xe_vm * vm) xe_vm_userptr_pin() argument 772 xe_vm_userptr_check_repin(struct xe_vm * vm) xe_vm_userptr_check_repin() argument 853 xe_vm_rebind(struct xe_vm * vm,bool rebind_worker) xe_vm_rebind() argument 908 xe_vma_rebind(struct xe_vm * vm,struct xe_vma * vma,u8 tile_mask) xe_vma_rebind() argument 962 xe_vma_create(struct xe_vm * vm,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 start,u64 end,u16 pat_index,unsigned int flags) xe_vma_create() argument 1066 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_destroy_late() local 1116 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_destroy() local 1160 struct xe_vm *vm = xe_vma_vm(vma); xe_vm_lock_vma() local 1192 xe_vm_find_overlapping_vma(struct xe_vm * vm,u64 start,u64 range) xe_vm_find_overlapping_vma() argument 1208 xe_vm_insert_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_insert_vma() argument 1223 xe_vm_remove_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_remove_vma() argument 1397 xe_vm_create_scratch(struct xe_device * xe,struct xe_tile * tile,struct xe_vm * vm) xe_vm_create_scratch() argument 1414 xe_vm_free_scratch(struct xe_vm * vm) xe_vm_free_scratch() argument 1437 struct xe_vm *vm; xe_vm_create() local 1583 xe_vm_close(struct xe_vm * vm) xe_vm_close() argument 1590 xe_vm_close_and_put(struct xe_vm * vm) xe_vm_close_and_put() argument 1694 struct xe_vm *vm = vm_destroy_work_func() local 1726 struct xe_vm *vm = container_of(gpuvm, struct xe_vm, gpuvm); xe_vm_free() local 1734 struct xe_vm *vm; xe_vm_lookup() local 1745 xe_vm_pdp4_descriptor(struct xe_vm * vm,struct xe_tile * tile) xe_vm_pdp4_descriptor() argument 1752 to_wait_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) to_wait_exec_queue() argument 1783 struct xe_vm *vm; xe_vm_create_ioctl() local 1868 struct xe_vm *vm; xe_vm_destroy_ioctl() local 1897 prep_vma_destroy(struct xe_vm * vm,struct xe_vma * vma,bool post_commit) prep_vma_destroy() argument 1962 vm_bind_ioctl_ops_create(struct xe_vm * vm,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 addr,u64 range,u32 operation,u32 flags,u32 prefetch_region,u16 pat_index) vm_bind_ioctl_ops_create() argument 2038 new_vma(struct xe_vm * vm,struct drm_gpuva_op_map * op,u16 pat_index,unsigned int flags) new_vma() argument 2122 xe_vma_op_commit(struct xe_vm * vm,struct xe_vma_op * op) xe_vma_op_commit() argument 2185 vm_bind_ioctl_ops_parse(struct xe_vm * vm,struct drm_gpuva_ops * ops,struct xe_vma_ops * vops) vm_bind_ioctl_ops_parse() argument 2342 xe_vma_op_unwind(struct xe_vm * vm,struct xe_vma_op * op,bool post_commit,bool prev_post_commit,bool next_post_commit) xe_vma_op_unwind() argument 2397 vm_bind_ioctl_ops_unwind(struct xe_vm * vm,struct drm_gpuva_ops ** ops,int num_ops_list) vm_bind_ioctl_ops_unwind() argument 2425 struct xe_vm *vm = xe_vma_vm(vma); vma_lock_and_validate() local 2454 op_lock_and_prep(struct drm_exec * exec,struct xe_vm * vm,struct xe_vma_op * op) op_lock_and_prep() argument 2510 vm_bind_ioctl_ops_lock_and_prep(struct drm_exec * exec,struct xe_vm * vm,struct xe_vma_ops * vops) vm_bind_ioctl_ops_lock_and_prep() argument 2567 vm_ops_setup_tile_args(struct xe_vm * vm,struct xe_vma_ops * vops) vm_ops_setup_tile_args() argument 2593 ops_execute(struct xe_vm * vm,struct xe_vma_ops * vops) ops_execute() argument 2687 op_add_ufence(struct xe_vm * vm,struct xe_vma_op * op,struct xe_user_fence * ufence) op_add_ufence() argument 2710 vm_bind_ioctl_ops_fini(struct xe_vm * vm,struct xe_vma_ops * vops,struct dma_fence * fence) vm_bind_ioctl_ops_fini() argument 2737 vm_bind_ioctl_ops_execute(struct xe_vm * vm,struct xe_vma_ops * vops) vm_bind_ioctl_ops_execute() argument 2899 vm_bind_ioctl_signal_fences(struct xe_vm * vm,struct xe_exec_queue * q,struct xe_sync_entry * syncs,int num_syncs) vm_bind_ioctl_signal_fences() argument 2922 xe_vma_ops_init(struct xe_vma_ops * vops,struct xe_vm * vm,struct xe_exec_queue * q,struct xe_sync_entry * syncs,u32 num_syncs) xe_vma_ops_init() argument 2991 struct xe_vm *vm; xe_vm_bind_ioctl() local 3204 xe_vm_lock(struct xe_vm * vm,bool intr) xe_vm_lock() argument 3218 xe_vm_unlock(struct xe_vm * vm) xe_vm_unlock() argument 3312 xe_vm_snapshot_capture(struct xe_vm * vm) xe_vm_snapshot_capture() argument [all...] |
H A D | xe_exec_queue.c | 51 __xe_exec_queue_alloc(struct xe_device * xe,struct xe_vm * vm,u32 logical_mask,u16 width,struct xe_hw_engine * hwe,u32 flags,u64 extensions) __xe_exec_queue_alloc() argument 113 struct xe_vm *vm = q->vm; __xe_exec_queue_init() local 148 xe_exec_queue_create(struct xe_device * xe,struct xe_vm * vm,u32 logical_mask,u16 width,struct xe_hw_engine * hwe,u32 flags,u64 extensions) xe_exec_queue_create() argument 173 xe_exec_queue_create_class(struct xe_device * xe,struct xe_gt * gt,struct xe_vm * vm,enum xe_engine_class class,u32 flags,u64 extensions) xe_exec_queue_create_class() argument 547 struct xe_vm *vm; xe_exec_queue_create_ioctl() local 858 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_lockdep_assert() argument 873 xe_exec_queue_last_fence_put(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_put() argument 904 xe_exec_queue_last_fence_get(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get() argument 931 xe_exec_queue_last_fence_get_for_resume(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get_for_resume() argument 955 xe_exec_queue_last_fence_set(struct xe_exec_queue * q,struct xe_vm * vm,struct dma_fence * fence) xe_exec_queue_last_fence_set() argument 972 xe_exec_queue_last_fence_test_dep(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_test_dep() argument [all...] |
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 122 struct amdgpu_vm *vm; member 141 int amdgpu_vm_set_pasid(struct amdgpu_device *adev, struct amdgpu_vm *vm, in amdgpu_vm_set_pasid() argument 180 struct amdgpu_vm *vm = vm_bo->vm; amdgpu_vm_bo_evicted() local 294 amdgpu_vm_bo_reset_state_machine(struct amdgpu_vm * vm) amdgpu_vm_bo_reset_state_machine() argument 324 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_shared() local 373 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_stats_locked() local 412 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_stats() local 430 amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base * base,struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_base_init() argument 479 amdgpu_vm_lock_pd(struct amdgpu_vm * vm,struct drm_exec * exec,unsigned int num_fences) amdgpu_vm_lock_pd() argument 497 amdgpu_vm_move_to_lru_tail(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_move_to_lru_tail() argument 506 amdgpu_vm_init_entities(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_init_entities() argument 526 amdgpu_vm_fini_entities(struct amdgpu_vm * vm) amdgpu_vm_fini_entities() argument 541 amdgpu_vm_generation(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_generation() argument 572 amdgpu_vm_validate(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket,int (* validate)(void * p,struct amdgpu_bo * bo),void * param) amdgpu_vm_validate() argument 659 amdgpu_vm_ready(struct amdgpu_vm * vm) amdgpu_vm_ready() argument 875 amdgpu_vm_bo_find(struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_find() argument 929 amdgpu_vm_update_pdes(struct amdgpu_device * adev,struct amdgpu_vm * vm,bool immediate) amdgpu_vm_update_pdes() argument 1014 struct amdgpu_vm *vm = params->vm; amdgpu_vm_tlb_flush() local 1064 amdgpu_vm_update_range(struct amdgpu_device * adev,struct amdgpu_vm * vm,bool immediate,bool unlocked,bool flush_tlb,bool allow_override,struct amdgpu_sync * sync,uint64_t start,uint64_t last,uint64_t flags,uint64_t offset,uint64_t vram_base,struct ttm_resource * res,dma_addr_t * pages_addr,struct dma_fence ** fence) amdgpu_vm_update_range() argument 1197 amdgpu_vm_get_memory(struct amdgpu_vm * vm,struct amdgpu_mem_stats stats[__AMDGPU_PL_NUM]) amdgpu_vm_get_memory() argument 1221 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_update() local 1457 amdgpu_vm_free_mapping(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_bo_va_mapping * mapping,struct dma_fence * fence) amdgpu_vm_free_mapping() argument 1474 amdgpu_vm_prt_fini(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_prt_fini() argument 1503 amdgpu_vm_clear_freed(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct dma_fence ** fence) amdgpu_vm_clear_freed() argument 1565 amdgpu_vm_handle_moved(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket) amdgpu_vm_handle_moved() argument 1643 amdgpu_vm_flush_compute_tlb(struct amdgpu_device * adev,struct amdgpu_vm * vm,uint32_t flush_type,uint32_t xcc_mask) amdgpu_vm_flush_compute_tlb() argument 1690 amdgpu_vm_bo_add(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_add() argument 1733 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_insert_map() local 1804 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_map() local 1913 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_unmap() local 1963 amdgpu_vm_bo_clear_mappings(struct amdgpu_device * adev,struct amdgpu_vm * vm,uint64_t saddr,uint64_t size) amdgpu_vm_bo_clear_mappings() argument 2081 amdgpu_vm_bo_lookup_mapping(struct amdgpu_vm * vm,uint64_t addr) amdgpu_vm_bo_lookup_mapping() argument 2095 amdgpu_vm_bo_trace_cs(struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket) amdgpu_vm_bo_trace_cs() argument 2132 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_del() local 2226 struct amdgpu_vm *vm = bo_base->vm; amdgpu_vm_bo_invalidate() local 2261 struct amdgpu_vm *vm = bo_base->vm; amdgpu_vm_bo_move() local 2393 amdgpu_vm_wait_idle(struct amdgpu_vm * vm,long timeout) amdgpu_vm_wait_idle() argument 2414 struct amdgpu_vm *vm; amdgpu_vm_get_vm_from_pasid() local 2445 amdgpu_vm_get_task_info_vm(struct amdgpu_vm * vm) amdgpu_vm_get_task_info_vm() argument 2473 amdgpu_vm_create_task_info(struct amdgpu_vm * vm) amdgpu_vm_create_task_info() argument 2488 amdgpu_vm_set_task_info(struct amdgpu_vm * vm) amdgpu_vm_set_task_info() argument 2518 amdgpu_vm_init(struct amdgpu_device * adev,struct amdgpu_vm * vm,int32_t xcp_id) amdgpu_vm_init() argument 2635 amdgpu_vm_make_compute(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_make_compute() argument 2685 amdgpu_vm_release_compute(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_release_compute() argument 2691 amdgpu_vm_stats_is_zero(struct amdgpu_vm * vm) amdgpu_vm_stats_is_zero() argument 2710 amdgpu_vm_fini(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_fini() argument 2913 struct amdgpu_vm *vm; amdgpu_vm_handle_fault() local 3003 amdgpu_debugfs_vm_bo_info(struct amdgpu_vm * vm,struct seq_file * m) amdgpu_debugfs_vm_bo_info() argument 3106 struct amdgpu_vm *vm; amdgpu_vm_update_fault_cache() local 3156 amdgpu_vm_is_bo_always_valid(struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_is_bo_always_valid() argument [all...] |
H A D | amdgpu_vm_pt.c | 157 struct amdgpu_vm *vm, uint64_t start, in amdgpu_vm_pt_start() argument 286 struct amdgpu_vm *vm, in amdgpu_vm_pt_first_dfs() argument 340 #define for_each_amdgpu_vm_pt_dfs_safe(adev, vm, start, cursor, entry) \ argument 359 int amdgpu_vm_pt_clear(struct amdgpu_device *adev, struct amdgpu_vm *vm, in amdgpu_vm_pt_clear() argument 438 amdgpu_vm_pt_create(struct amdgpu_device * adev,struct amdgpu_vm * vm,int level,bool immediate,struct amdgpu_bo_vm ** vmbo,int32_t xcp_id) amdgpu_vm_pt_create() argument 494 amdgpu_vm_pt_alloc(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_vm_pt_cursor * cursor,bool immediate) amdgpu_vm_pt_alloc() argument 553 struct amdgpu_vm *vm; amdgpu_vm_pt_free_work() local 583 struct amdgpu_vm *vm = params->vm; amdgpu_vm_pt_free_list() local 633 amdgpu_vm_pt_free_root(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_pt_free_root() argument 657 struct amdgpu_vm *vm = params->vm; amdgpu_vm_pde_update() local 908 struct amdgpu_vm *vm = params->vm; amdgpu_vm_ptes_update() local 972 amdgpu_vm_pt_map_tables(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_pt_map_tables() argument [all...] |
/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_gtt.c | 41 struct drm_i915_gem_object *alloc_pt_lmem(struct i915_address_space *vm, int sz) in alloc_pt_lmem() 75 struct drm_i915_gem_object *alloc_pt_dma(struct i915_address_space *vm, int sz) in alloc_pt_dma() 99 int map_pt_dma(struct i915_address_space *vm, struct drm_i915_gem_object *obj) in map_pt_dma() 123 int map_pt_dma_locked(struct i915_address_space *vm, struct drm_i915_gem_object *obj) in map_pt_dma_locked() 183 static void __i915_vm_close(struct i915_address_space *vm) in __i915_vm_close() 198 int i915_vm_lock_objects(struct i915_address_space *vm, in i915_vm_lock_objects() 211 void i915_address_space_fini(struct i915_address_space *vm) in i915_address_space_fini() 226 struct i915_address_space *vm = in i915_vm_resv_release() local 237 struct i915_address_space *vm = in __i915_vm_release() local 253 struct i915_address_space *vm = in i915_vm_release() local [all …]
|
H A D | gen8_ppgtt.c | 180 static unsigned int gen8_pd_top_count(const struct i915_address_space *vm) in gen8_pd_top_count() 188 gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) in gen8_pdp_for_page_index() 199 gen8_pdp_for_page_address(struct i915_address_space * const vm, const u64 addr) in gen8_pdp_for_page_address() 204 static void __gen8_ppgtt_cleanup(struct i915_address_space *vm, in __gen8_ppgtt_cleanup() 222 static void gen8_ppgtt_cleanup(struct i915_address_space *vm) in gen8_ppgtt_cleanup() 239 static u64 __gen8_ppgtt_clear(struct i915_address_space * const vm, in __gen8_ppgtt_clear() 307 static void gen8_ppgtt_clear(struct i915_address_space *vm, in gen8_ppgtt_clear() 322 static void __gen8_ppgtt_alloc(struct i915_address_space * const vm, in __gen8_ppgtt_alloc() 390 static void gen8_ppgtt_alloc(struct i915_address_space *vm, in gen8_ppgtt_alloc() 406 static void __gen8_ppgtt_foreach(struct i915_address_space *vm, in __gen8_ppgtt_foreach() [all …]
|
H A D | intel_ggtt.c | 115 void i915_ggtt_suspend_vm(struct i915_address_space *vm, bool evict_all) in i915_ggtt_suspend_vm() 438 static void gen8_ggtt_insert_page(struct i915_address_space *vm, in gen8_ggtt_insert_page() 453 static void gen8_ggtt_insert_page_bind(struct i915_address_space *vm, in gen8_ggtt_insert_page_bind() 468 static void gen8_ggtt_insert_entries(struct i915_address_space *vm, in gen8_ggtt_insert_entries() 507 static bool __gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, in __gen8_ggtt_insert_entries_bind() 538 static void gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, in gen8_ggtt_insert_entries_bind() 551 static void gen8_ggtt_clear_range(struct i915_address_space *vm, in gen8_ggtt_clear_range() 572 static void gen8_ggtt_scratch_range_bind(struct i915_address_space *vm, in gen8_ggtt_scratch_range_bind() 593 static void gen6_ggtt_insert_page(struct i915_address_space *vm, in gen6_ggtt_insert_page() 614 static void gen6_ggtt_insert_entries(struct i915_address_space *vm, in gen6_ggtt_insert_entries() [all …]
|
H A D | intel_ppgtt.c | 16 struct i915_page_table *alloc_pt(struct i915_address_space *vm, int sz) in alloc_pt() 53 struct i915_page_directory *alloc_pd(struct i915_address_space *vm) in alloc_pd() 71 void free_px(struct i915_address_space *vm, struct i915_page_table *pt, int lvl) in free_px() 182 void ppgtt_bind_vma(struct i915_address_space *vm, in ppgtt_bind_vma() 207 void ppgtt_unbind_vma(struct i915_address_space *vm, in ppgtt_unbind_vma() 223 int i915_vm_alloc_pt_stash(struct i915_address_space *vm, in i915_vm_alloc_pt_stash() 276 int i915_vm_map_pt_stash(struct i915_address_space *vm, in i915_vm_map_pt_stash() 293 void i915_vm_free_pt_stash(struct i915_address_space *vm, in i915_vm_free_pt_stash()
|
H A D | intel_gtt.h | 367 struct i915_address_space vm; member 410 struct i915_address_space vm; member 415 #define i915_is_ggtt(vm) ((vm)->is_ggtt) argument 416 #define i915_is_dpt(vm) ((vm)->is_dpt) argument 417 #define i915_is_ggtt_or_dpt(vm) (i915_is_ggtt(vm) || i915_is_dpt(vm)) argument 425 i915_vm_is_4lvl(const struct i915_address_space *vm) in i915_vm_is_4lvl() 431 i915_vm_has_scratch_64K(struct i915_address_space *vm) in i915_vm_has_scratch_64K() 436 static inline u64 i915_vm_min_alignment(struct i915_address_space *vm, in i915_vm_min_alignment() 446 static inline u64 i915_vm_obj_min_alignment(struct i915_address_space *vm, in i915_vm_obj_min_alignment() 456 i915_vm_has_cache_coloring(struct i915_address_space *vm) in i915_vm_has_cache_coloring() [all …]
|
H A D | intel_ggtt_gmch.c | 18 static void gmch_ggtt_insert_page(struct i915_address_space *vm, in gmch_ggtt_insert_page() 30 static void gmch_ggtt_insert_entries(struct i915_address_space *vm, in gmch_ggtt_insert_entries() 47 static void gmch_ggtt_clear_range(struct i915_address_space *vm, in gmch_ggtt_clear_range() 53 static void gmch_ggtt_remove(struct i915_address_space *vm) in gmch_ggtt_remove()
|
/linux/drivers/gpu/drm/panthor/ |
H A D | panthor_mmu.c | 46 struct panthor_vm *vm; member 405 /** @vm member 101 } vm; global() member 437 struct panthor_vm *vm = cookie; alloc_pt() local 489 struct panthor_vm *vm = cookie; free_pt() local 595 mmu_hw_do_operation(struct panthor_vm * vm,u64 iova,u64 size,u32 op) mmu_hw_do_operation() argument 666 panthor_vm_has_unhandled_faults(struct panthor_vm * vm) panthor_vm_has_unhandled_faults() argument 677 panthor_vm_is_unusable(struct panthor_vm * vm) panthor_vm_is_unusable() argument 682 panthor_vm_release_as_locked(struct panthor_vm * vm) panthor_vm_release_as_locked() argument 706 panthor_vm_active(struct panthor_vm * vm) panthor_vm_active() argument 815 panthor_vm_idle(struct panthor_vm * vm) panthor_vm_idle() argument 829 panthor_vm_page_size(struct panthor_vm * vm) panthor_vm_page_size() argument 837 panthor_vm_stop(struct panthor_vm * vm) panthor_vm_stop() argument 842 panthor_vm_start(struct panthor_vm * vm) panthor_vm_start() argument 853 panthor_vm_as(struct panthor_vm * vm) panthor_vm_as() argument 879 panthor_vm_flush_range(struct panthor_vm * vm,u64 iova,u64 size) panthor_vm_flush_range() argument 903 panthor_vm_flush_all(struct panthor_vm * vm) panthor_vm_flush_all() argument 908 panthor_vm_unmap_pages(struct panthor_vm * vm,u64 iova,u64 size) panthor_vm_unmap_pages() argument 937 panthor_vm_map_pages(struct panthor_vm * vm,u64 iova,int prot,struct sg_table * sgt,u64 offset,u64 size) panthor_vm_map_pages() argument 1035 panthor_vm_alloc_va(struct panthor_vm * vm,u64 va,u64 size,struct drm_mm_node * va_node) panthor_vm_alloc_va() argument 1069 panthor_vm_free_va(struct panthor_vm * vm,struct drm_mm_node * va_node) panthor_vm_free_va() argument 1079 struct drm_gpuvm *vm = vm_bo->vm; panthor_vm_bo_put() local 1113 panthor_vm_cleanup_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm) panthor_vm_cleanup_op_ctx() argument 1199 panthor_vm_prepare_map_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm,struct panthor_gem_object * bo,u64 offset,u64 size,u64 va,u32 flags) panthor_vm_prepare_map_op_ctx() argument 1327 panthor_vm_prepare_unmap_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm,u64 va,u64 size) panthor_vm_prepare_unmap_op_ctx() argument 1379 panthor_vm_prepare_sync_only_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm) panthor_vm_prepare_sync_only_op_ctx() argument 1401 panthor_vm_get_bo_for_va(struct panthor_vm * vm,u64 va,u64 * bo_offset) panthor_vm_get_bo_for_va() argument 1506 struct panthor_vm *vm; panthor_vm_pool_create_vm() local 1531 panthor_vm_destroy(struct panthor_vm * vm) panthor_vm_destroy() argument 1564 struct panthor_vm *vm; panthor_vm_pool_destroy_vm() local 1583 struct panthor_vm *vm; panthor_vm_pool_get_vm() local 1603 struct panthor_vm *vm; panthor_vm_pool_destroy() local 1745 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_suspend() local 1788 struct panthor_vm *vm; panthor_mmu_pre_reset() local 1808 struct panthor_vm *vm; panthor_mmu_post_reset() local 1819 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_post_reset() local 1840 struct panthor_vm *vm = container_of(gpuvm, struct panthor_vm, base); panthor_vm_free() local 1888 panthor_vm_put(struct panthor_vm * vm) panthor_vm_put() argument 1899 panthor_vm_get(struct panthor_vm * vm) panthor_vm_get() argument 1921 panthor_vm_get_heap_pool(struct panthor_vm * vm,bool create) panthor_vm_get_heap_pool() argument 1986 panthor_vma_link(struct panthor_vm * vm,struct panthor_vma * vma,struct drm_gpuvm_bo * vm_bo) panthor_vma_link() argument 1998 panthor_vma_unlink(struct panthor_vm * vm,struct panthor_vma * vma) panthor_vma_unlink() argument 2029 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_map() local 2058 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_remap() local 2106 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_unmap() local 2132 panthor_vm_resv(struct panthor_vm * vm) panthor_vm_resv() argument 2137 panthor_vm_root_gem(struct panthor_vm * vm) panthor_vm_root_gem() argument 2146 panthor_vm_exec_op(struct panthor_vm * vm,struct panthor_vm_op_ctx * op,bool flag_vm_unusable_on_failure) panthor_vm_exec_op() argument 2280 struct panthor_vm *vm; panthor_vm_create() local 2385 panthor_vm_bind_prepare_op_ctx(struct drm_file * file,struct panthor_vm * vm,const struct drm_panthor_vm_bind_op * op,struct panthor_vm_op_ctx * op_ctx) panthor_vm_bind_prepare_op_ctx() argument 2457 panthor_vm_bind_job_create(struct drm_file * file,struct panthor_vm * vm,const struct drm_panthor_vm_bind_op * op) panthor_vm_bind_job_create() argument 2543 panthor_vm_update_resvs(struct panthor_vm * vm,struct drm_exec * exec,struct dma_fence * fence,enum dma_resv_usage private_usage,enum dma_resv_usage extobj_usage) panthor_vm_update_resvs() argument 2560 panthor_vm_bind_exec_sync_op(struct drm_file * file,struct panthor_vm * vm,struct drm_panthor_vm_bind_op * op) panthor_vm_bind_exec_sync_op() argument 2598 panthor_vm_map_bo_range(struct panthor_vm * vm,struct panthor_gem_object * bo,u64 offset,u64 size,u64 va,u32 flags) panthor_vm_map_bo_range() argument 2625 panthor_vm_unmap_range(struct panthor_vm * vm,u64 va,u64 size) panthor_vm_unmap_range() argument 2653 panthor_vm_prepare_mapped_bos_resvs(struct drm_exec * exec,struct panthor_vm * vm,u32 slot_count) panthor_vm_prepare_mapped_bos_resvs() argument 2680 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_unplug() local 2750 show_vm_gpuvas(struct panthor_vm * vm,struct seq_file * m) show_vm_gpuvas() argument 2767 struct panthor_vm *vm; show_each_vm() local [all...] |
/linux/drivers/gpu/drm/i915/display/ |
H A D | intel_dpt.c | 17 struct i915_address_space vm; member 24 #define i915_is_dpt(vm) ((vm)->is_dpt) argument 27 i915_vm_to_dpt(struct i915_address_space *vm) in i915_vm_to_dpt() argument 41 dpt_insert_page(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) dpt_insert_page() argument 54 dpt_insert_entries(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) dpt_insert_entries() argument 76 dpt_clear_range(struct i915_address_space * vm,u64 start,u64 length) dpt_clear_range() argument 81 dpt_bind_vma(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) dpt_bind_vma() argument 111 dpt_unbind_vma(struct i915_address_space * vm,struct i915_vma_resource * vma_res) dpt_unbind_vma() argument 117 dpt_cleanup(struct i915_address_space * vm) dpt_cleanup() argument 124 intel_dpt_pin_to_ggtt(struct i915_address_space * vm,unsigned int alignment) intel_dpt_pin_to_ggtt() argument 176 intel_dpt_unpin_from_ggtt(struct i915_address_space * vm) intel_dpt_unpin_from_ggtt() argument 248 struct i915_address_space *vm; intel_dpt_create() local 312 intel_dpt_destroy(struct i915_address_space * vm) intel_dpt_destroy() argument [all...] |
/linux/drivers/gpu/drm/i915/selftests/ |
H A D | mock_gtt.c | 27 static void mock_insert_page(struct i915_address_space *vm, in mock_insert_page() 35 static void mock_insert_entries(struct i915_address_space *vm, in mock_insert_entries() 41 static void mock_bind_ppgtt(struct i915_address_space *vm, in mock_bind_ppgtt() 51 static void mock_unbind_ppgtt(struct i915_address_space *vm, in mock_unbind_ppgtt() 56 static void mock_cleanup(struct i915_address_space *vm) in mock_cleanup() 60 static void mock_clear_range(struct i915_address_space *vm, in mock_clear_range() 94 static void mock_bind_ggtt(struct i915_address_space *vm, in mock_bind_ggtt() 102 static void mock_unbind_ggtt(struct i915_address_space *vm, in mock_unbind_ggtt()
|
H A D | i915_gem_gtt.c | 244 lowlevel_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) lowlevel_hole() argument 391 close_object_list(struct list_head * objects,struct i915_address_space * vm) close_object_list() argument 408 fill_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) fill_hole() argument 633 walk_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) walk_hole() argument 716 pot_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) pot_hole() argument 791 drunk_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) drunk_hole() argument 901 __shrink_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) __shrink_hole() argument 979 shrink_hole(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) shrink_hole() argument 1001 shrink_boom(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) shrink_boom() argument 1079 misaligned_case(struct i915_address_space * vm,struct intel_memory_region * mr,u64 addr,u64 size,unsigned long flags) misaligned_case() argument 1144 misaligned_pin(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time) misaligned_pin() argument 1191 exercise_ppgtt(struct drm_i915_private * dev_priv,int (* func)(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time)) exercise_ppgtt() argument 1277 exercise_ggtt(struct drm_i915_private * i915,int (* func)(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time)) exercise_ggtt() argument 1454 exercise_mock(struct drm_i915_private * i915,int (* func)(struct i915_address_space * vm,u64 hole_start,u64 hole_end,unsigned long end_time)) exercise_mock() argument 1459 struct i915_address_space *vm; exercise_mock() local 1506 struct i915_address_space *vm = vma->vm; reserve_gtt_with_resource() local 1687 struct i915_address_space *vm = vma->vm; insert_gtt_with_resource() local [all...] |
/linux/drivers/virt/acrn/ |
H A D | irqfd.c | 32 struct acrn_vm *vm; member 43 struct acrn_vm *vm = irqfd->vm; in acrn_irqfd_inject() local 65 struct acrn_vm *vm; in hsm_irqfd_shutdown_work() local 81 struct acrn_vm *vm; in hsm_irqfd_wakeup() local 110 static int acrn_irqfd_assign(struct acrn_vm *vm, struct acrn_irqfd *args) in acrn_irqfd_assign() 172 static int acrn_irqfd_deassign(struct acrn_vm *vm, in acrn_irqfd_deassign() 195 int acrn_irqfd_config(struct acrn_vm *vm, struct acrn_irqfd *args) in acrn_irqfd_config() 207 int acrn_irqfd_init(struct acrn_vm *vm) in acrn_irqfd_init() 219 void acrn_irqfd_deinit(struct acrn_vm *vm) in acrn_irqfd_deinit()
|
H A D | ioeventfd.c | 43 static void acrn_ioeventfd_shutdown(struct acrn_vm *vm, struct hsm_ioeventfd *p) in acrn_ioeventfd_shutdown() 52 static bool hsm_ioeventfd_is_conflict(struct acrn_vm *vm, in hsm_ioeventfd_is_conflict() 76 static int acrn_ioeventfd_assign(struct acrn_vm *vm, in acrn_ioeventfd_assign() 147 static int acrn_ioeventfd_deassign(struct acrn_vm *vm, in acrn_ioeventfd_deassign() 173 static struct hsm_ioeventfd *hsm_ioeventfd_match(struct acrn_vm *vm, u64 addr, in hsm_ioeventfd_match() 232 int acrn_ioeventfd_config(struct acrn_vm *vm, struct acrn_ioeventfd *args) in acrn_ioeventfd_config() 244 int acrn_ioeventfd_init(struct acrn_vm *vm) in acrn_ioeventfd_init() 263 void acrn_ioeventfd_deinit(struct acrn_vm *vm) in acrn_ioeventfd_deinit()
|
H A D | ioreq.c | 39 static int ioreq_complete_request(struct acrn_vm *vm, u16 vcpu, in ioreq_complete_request() 93 int acrn_ioreq_request_default_complete(struct acrn_vm *vm, u16 vcpu) in acrn_ioreq_request_default_complete() 212 void acrn_ioreq_request_clear(struct acrn_vm *vm) in acrn_ioreq_request_clear() 312 static bool handle_cf8cfc(struct acrn_vm *vm, in handle_cf8cfc() 381 static struct acrn_ioreq_client *find_ioreq_client(struct acrn_vm *vm, in find_ioreq_client() 415 struct acrn_ioreq_client *acrn_ioreq_client_create(struct acrn_vm *vm, in acrn_ioreq_client_create() 468 struct acrn_vm *vm = client->vm; in acrn_ioreq_client_destroy() local 497 static int acrn_ioreq_dispatch(struct acrn_vm *vm) in acrn_ioreq_dispatch() 546 struct acrn_vm *vm; in ioreq_dispatcher() local 596 int acrn_ioreq_init(struct acrn_vm *vm, u64 buf_vma) in acrn_ioreq_init() [all …]
|
H A D | mm.c | 19 static int modify_region(struct acrn_vm *vm, struct vm_memory_region_op *region) in modify_region() 52 int acrn_mm_region_add(struct acrn_vm *vm, u64 user_gpa, u64 service_gpa, in acrn_mm_region_add() 85 int acrn_mm_region_del(struct acrn_vm *vm, u64 user_gpa, u64 size) in acrn_mm_region_del() 108 int acrn_vm_memseg_map(struct acrn_vm *vm, struct acrn_vm_memmap *memmap) in acrn_vm_memseg_map() 131 int acrn_vm_memseg_unmap(struct acrn_vm *vm, struct acrn_vm_memmap *memmap) in acrn_vm_memseg_unmap() 156 int acrn_vm_ram_map(struct acrn_vm *vm, struct acrn_vm_memmap *memmap) in acrn_vm_ram_map() 354 void acrn_vm_all_ram_unmap(struct acrn_vm *vm) in acrn_vm_all_ram_unmap()
|
/linux/drivers/gpu/drm/lima/ |
H A D | lima_vm.c | 18 struct lima_vm *vm; member 35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) in lima_vm_unmap_range() 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() 77 lima_vm_bo_find(struct lima_vm *vm, struct lima_bo *bo) in lima_vm_bo_find() 91 int lima_vm_bo_add(struct lima_vm *vm, struct lima_bo *bo, bool create) in lima_vm_bo_add() 155 void lima_vm_bo_del(struct lima_vm *vm, struct lima_bo *bo) in lima_vm_bo_del() 185 u32 lima_vm_get_va(struct lima_vm *vm, struct lima_bo *bo) in lima_vm_get_va() 202 struct lima_vm *vm; in lima_vm_create() local 237 struct lima_vm *vm = container_of(kref, struct lima_vm, refcount); in lima_vm_release() local 254 void lima_vm_print(struct lima_vm *vm) in lima_vm_print() [all …]
|
/linux/sound/pci/ctxfi/ |
H A D | ctvmem.c | 30 get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc) in get_vm_block() 75 static void put_vm_block(struct ct_vm *vm, struct ct_vm_block *block) in put_vm_block() 124 ct_vm_map(struct ct_vm *vm, struct snd_pcm_substream *substream, int size) in ct_vm_map() 152 static void ct_vm_unmap(struct ct_vm *vm, struct ct_vm_block *block) in ct_vm_unmap() 164 ct_get_ptp_phys(struct ct_vm *vm, int index) in ct_get_ptp_phys() 171 struct ct_vm *vm; in ct_vm_create() local 215 void ct_vm_destroy(struct ct_vm *vm) in ct_vm_destroy()
|
/linux/tools/testing/selftests/kvm/lib/ |
H A D | memstress.c | 88 void memstress_setup_vcpus(struct kvm_vm *vm, int nr_vcpus, in memstress_setup_vcpus() 130 struct kvm_vm *vm; in memstress_create_vm() local 234 void memstress_destroy_vm(struct kvm_vm *vm) in memstress_destroy_vm() 239 void memstress_set_write_percent(struct kvm_vm *vm, uint32_t write_percent) in memstress_set_write_percent() 245 void memstress_set_random_access(struct kvm_vm *vm, bool random_access) in memstress_set_random_access() 256 void __weak memstress_setup_nested(struct kvm_vm *vm, int nr_vcpus, struct kvm_vcpu **vcpus) in memstress_setup_nested() 322 static void toggle_dirty_logging(struct kvm_vm *vm, int slots, bool enable) in toggle_dirty_logging() 334 void memstress_enable_dirty_logging(struct kvm_vm *vm, int slots) in memstress_enable_dirty_logging() 339 void memstress_disable_dirty_logging(struct kvm_vm *vm, int slots) in memstress_disable_dirty_logging() 344 void memstress_get_dirty_log(struct kvm_vm *vm, unsigned long *bitmaps[], int slots) in memstress_get_dirty_log() [all …]
|
/linux/drivers/gpu/drm/radeon/ |
H A D | radeon_vm.c | 130 struct radeon_vm *vm, in radeon_vm_get_bos() 177 struct radeon_vm *vm, int ring) in radeon_vm_grab_id() 236 struct radeon_vm *vm, in radeon_vm_flush() 268 struct radeon_vm *vm, in radeon_vm_fence() 292 struct radeon_bo_va *radeon_vm_bo_find(struct radeon_vm *vm, in radeon_vm_bo_find() 319 struct radeon_vm *vm, in radeon_vm_bo_add() 451 struct radeon_vm *vm = bo_va->vm; in radeon_vm_bo_set_addr() local 639 struct radeon_vm *vm) in radeon_vm_update_page_directory() 813 struct radeon_vm *vm, in radeon_vm_update_ptes() 884 static void radeon_vm_fence_pts(struct radeon_vm *vm, in radeon_vm_fence_pts() [all …]
|
/linux/drivers/gpu/drm/i915/ |
H A D | i915_vma_resource.c | 105 struct i915_address_space *vm; in __i915_vma_resource_unhold() local 177 struct i915_address_space *vm = vma_res->vm; in i915_vma_resource_unbind_work() local 233 struct i915_address_space *vm = vma_res->vm; in i915_vma_resource_unbind() local 272 i915_vma_resource_color_adjust_range(struct i915_address_space *vm, in i915_vma_resource_color_adjust_range() 295 int i915_vma_resource_bind_dep_sync(struct i915_address_space *vm, in i915_vma_resource_bind_dep_sync() 330 void i915_vma_resource_bind_dep_sync_all(struct i915_address_space *vm) in i915_vma_resource_bind_dep_sync_all() 379 int i915_vma_resource_bind_dep_await(struct i915_address_space *vm, in i915_vma_resource_bind_dep_await()
|