| /linux/drivers/virtio/ |
| H A D | virtio_mem.c | 285 static void virtio_mem_retry(struct virtio_mem *vm); 286 static int virtio_mem_create_resource(struct virtio_mem *vm); 287 static void virtio_mem_delete_resource(struct virtio_mem *vm); 293 static int register_virtio_mem_device(struct virtio_mem *vm) in register_virtio_mem_device() argument 302 list_add_rcu(&vm->next, &virtio_mem_devices); in register_virtio_mem_device() 312 static void unregister_virtio_mem_device(struct virtio_mem *vm) in unregister_virtio_mem_device() argument 316 list_del_rcu(&vm->next); in unregister_virtio_mem_device() 343 static unsigned long virtio_mem_phys_to_bb_id(struct virtio_mem *vm, in virtio_mem_phys_to_bb_id() argument 346 return addr / vm->bbm.bb_size; in virtio_mem_phys_to_bb_id() 352 static uint64_t virtio_mem_bb_id_to_phys(struct virtio_mem *vm, in virtio_mem_bb_id_to_phys() argument 361 virtio_mem_phys_to_sb_id(struct virtio_mem * vm,unsigned long addr) virtio_mem_phys_to_sb_id() argument 373 virtio_mem_bbm_set_bb_state(struct virtio_mem * vm,unsigned long bb_id,enum virtio_mem_bbm_bb_state state) virtio_mem_bbm_set_bb_state() argument 391 virtio_mem_bbm_get_bb_state(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_get_bb_state() argument 400 virtio_mem_bbm_bb_states_prepare_next_bb(struct virtio_mem * vm) virtio_mem_bbm_bb_states_prepare_next_bb() argument 440 virtio_mem_sbm_set_mb_state(struct virtio_mem * vm,unsigned long mb_id,uint8_t state) virtio_mem_sbm_set_mb_state() argument 457 virtio_mem_sbm_get_mb_state(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_get_mb_state() argument 468 virtio_mem_sbm_mb_states_prepare_next_mb(struct virtio_mem * vm) virtio_mem_sbm_mb_states_prepare_next_mb() argument 507 virtio_mem_sbm_sb_state_bit_nr(struct virtio_mem * vm,unsigned long mb_id,int sb_id) virtio_mem_sbm_sb_state_bit_nr() argument 518 virtio_mem_sbm_set_sb_plugged(struct virtio_mem * vm,unsigned long mb_id,int sb_id,int count) virtio_mem_sbm_set_sb_plugged() argument 532 virtio_mem_sbm_set_sb_unplugged(struct virtio_mem * vm,unsigned long mb_id,int sb_id,int count) virtio_mem_sbm_set_sb_unplugged() argument 544 virtio_mem_sbm_test_sb_plugged(struct virtio_mem * vm,unsigned long mb_id,int sb_id,int count) virtio_mem_sbm_test_sb_plugged() argument 561 virtio_mem_sbm_test_sb_unplugged(struct virtio_mem * vm,unsigned long mb_id,int sb_id,int count) virtio_mem_sbm_test_sb_unplugged() argument 576 virtio_mem_sbm_first_unplugged_sb(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_first_unplugged_sb() argument 588 virtio_mem_sbm_sb_states_prepare_next_mb(struct virtio_mem * vm) virtio_mem_sbm_sb_states_prepare_next_mb() argument 620 virtio_mem_could_add_memory(struct virtio_mem * vm,uint64_t size) virtio_mem_could_add_memory() argument 636 virtio_mem_add_memory(struct virtio_mem * vm,uint64_t addr,uint64_t size) virtio_mem_add_memory() argument 672 virtio_mem_sbm_add_mb(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_add_mb() argument 683 virtio_mem_bbm_add_bb(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_add_bb() argument 700 virtio_mem_remove_memory(struct virtio_mem * vm,uint64_t addr,uint64_t size) virtio_mem_remove_memory() argument 724 virtio_mem_sbm_remove_mb(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_remove_mb() argument 740 virtio_mem_offline_and_remove_memory(struct virtio_mem * vm,uint64_t addr,uint64_t size) virtio_mem_offline_and_remove_memory() argument 773 virtio_mem_sbm_offline_and_remove_mb(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_offline_and_remove_mb() argument 788 virtio_mem_sbm_try_remove_unplugged_mb(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_try_remove_unplugged_mb() argument 814 virtio_mem_bbm_offline_and_remove_bb(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_offline_and_remove_bb() argument 826 virtio_mem_retry(struct virtio_mem * vm) virtio_mem_retry() argument 836 virtio_mem_translate_node_id(struct virtio_mem * vm,uint16_t node_id) virtio_mem_translate_node_id() argument 851 virtio_mem_overlaps_range(struct virtio_mem * vm,uint64_t start,uint64_t size) virtio_mem_overlaps_range() argument 861 virtio_mem_contains_range(struct virtio_mem * vm,uint64_t start,uint64_t size) virtio_mem_contains_range() argument 867 virtio_mem_sbm_notify_going_online(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_notify_going_online() argument 882 virtio_mem_sbm_notify_offline(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_notify_offline() argument 902 virtio_mem_sbm_notify_online(struct virtio_mem * vm,unsigned long mb_id,unsigned long start_pfn) virtio_mem_sbm_notify_online() argument 927 virtio_mem_sbm_notify_going_offline(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_notify_going_offline() argument 943 virtio_mem_sbm_notify_cancel_offline(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_notify_cancel_offline() argument 959 virtio_mem_bbm_notify_going_offline(struct virtio_mem * vm,unsigned long bb_id,unsigned long pfn,unsigned long nr_pages) virtio_mem_bbm_notify_going_offline() argument 974 virtio_mem_bbm_notify_cancel_offline(struct virtio_mem * vm,unsigned long bb_id,unsigned long pfn,unsigned long nr_pages) virtio_mem_bbm_notify_cancel_offline() argument 993 struct virtio_mem *vm = container_of(nb, struct virtio_mem, virtio_mem_memory_notifier_cb() local 1124 struct virtio_mem *vm = container_of(nb, struct virtio_mem, virtio_mem_pm_notifier_cb() local 1224 virtio_mem_fake_offline(struct virtio_mem * vm,unsigned long pfn,unsigned long nr_pages) virtio_mem_fake_offline() argument 1299 virtio_mem_online_page(struct virtio_mem * vm,struct page * page,unsigned int order) virtio_mem_online_page() argument 1361 struct virtio_mem *vm; virtio_mem_online_page_cb() local 1389 virtio_mem_send_request(struct virtio_mem * vm,const struct virtio_mem_req * req) virtio_mem_send_request() argument 1419 virtio_mem_send_plug_request(struct virtio_mem * vm,uint64_t addr,uint64_t size) virtio_mem_send_plug_request() argument 1457 virtio_mem_send_unplug_request(struct virtio_mem * vm,uint64_t addr,uint64_t size) virtio_mem_send_unplug_request() argument 1492 virtio_mem_send_unplug_all_request(struct virtio_mem * vm) virtio_mem_send_unplug_all_request() argument 1523 virtio_mem_sbm_plug_sb(struct virtio_mem * vm,unsigned long mb_id,int sb_id,int count) virtio_mem_sbm_plug_sb() argument 1541 virtio_mem_sbm_unplug_sb(struct virtio_mem * vm,unsigned long mb_id,int sb_id,int count) virtio_mem_sbm_unplug_sb() argument 1560 virtio_mem_bbm_unplug_bb(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_unplug_bb() argument 1573 virtio_mem_bbm_plug_bb(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_plug_bb() argument 1590 virtio_mem_sbm_unplug_any_sb_raw(struct virtio_mem * vm,unsigned long mb_id,uint64_t * nb_sb) virtio_mem_sbm_unplug_any_sb_raw() argument 1629 virtio_mem_sbm_unplug_mb(struct virtio_mem * vm,unsigned long mb_id) virtio_mem_sbm_unplug_mb() argument 1639 virtio_mem_sbm_prepare_next_mb(struct virtio_mem * vm,unsigned long * mb_id) virtio_mem_sbm_prepare_next_mb() argument 1668 virtio_mem_sbm_plug_and_add_mb(struct virtio_mem * vm,unsigned long mb_id,uint64_t * nb_sb) virtio_mem_sbm_plug_and_add_mb() argument 1719 virtio_mem_sbm_plug_any_sb(struct virtio_mem * vm,unsigned long mb_id,uint64_t * nb_sb) virtio_mem_sbm_plug_any_sb() argument 1760 virtio_mem_sbm_plug_request(struct virtio_mem * vm,uint64_t diff) virtio_mem_sbm_plug_request() argument 1828 virtio_mem_bbm_plug_and_add_bb(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_plug_and_add_bb() argument 1859 virtio_mem_bbm_prepare_next_bb(struct virtio_mem * vm,unsigned long * bb_id) virtio_mem_bbm_prepare_next_bb() argument 1878 virtio_mem_bbm_plug_request(struct virtio_mem * vm,uint64_t diff) virtio_mem_bbm_plug_request() argument 1922 virtio_mem_plug_request(struct virtio_mem * vm,uint64_t diff) virtio_mem_plug_request() argument 1938 virtio_mem_sbm_unplug_any_sb_offline(struct virtio_mem * vm,unsigned long mb_id,uint64_t * nb_sb) virtio_mem_sbm_unplug_any_sb_offline() argument 1976 virtio_mem_sbm_unplug_sb_online(struct virtio_mem * vm,unsigned long mb_id,int sb_id,int count) virtio_mem_sbm_unplug_sb_online() argument 2024 virtio_mem_sbm_unplug_any_sb_online(struct virtio_mem * vm,unsigned long mb_id,uint64_t * nb_sb) virtio_mem_sbm_unplug_any_sb_online() argument 2079 virtio_mem_sbm_unplug_any_sb(struct virtio_mem * vm,unsigned long mb_id,uint64_t * nb_sb) virtio_mem_sbm_unplug_any_sb() argument 2098 virtio_mem_sbm_unplug_request(struct virtio_mem * vm,uint64_t diff) virtio_mem_sbm_unplug_request() argument 2158 virtio_mem_bbm_offline_remove_and_unplug_bb(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_offline_remove_and_unplug_bb() argument 2224 virtio_mem_bbm_bb_is_offline(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_bb_is_offline() argument 2243 virtio_mem_bbm_bb_is_movable(struct virtio_mem * vm,unsigned long bb_id) virtio_mem_bbm_bb_is_movable() argument 2263 virtio_mem_bbm_unplug_request(struct virtio_mem * vm,uint64_t diff) virtio_mem_bbm_unplug_request() argument 2306 virtio_mem_unplug_request(struct virtio_mem * vm,uint64_t diff) virtio_mem_unplug_request() argument 2318 virtio_mem_cleanup_pending_mb(struct virtio_mem * vm) virtio_mem_cleanup_pending_mb() argument 2370 virtio_mem_refresh_config(struct virtio_mem * vm) virtio_mem_refresh_config() argument 2417 struct virtio_mem *vm = container_of(work, struct virtio_mem, wq); virtio_mem_run_wq() local 2506 struct virtio_mem *vm = container_of(timer, struct virtio_mem, virtio_mem_timer_expired() local 2517 struct virtio_mem *vm = vq->vdev->priv; virtio_mem_handle_response() local 2522 virtio_mem_init_vq(struct virtio_mem * vm) virtio_mem_init_vq() argument 2535 virtio_mem_init_hotplug(struct virtio_mem * vm) virtio_mem_init_hotplug() argument 2667 virtio_mem_send_state_request(struct virtio_mem * vm,uint64_t addr,uint64_t size) virtio_mem_send_state_request() argument 2698 struct virtio_mem *vm = container_of(cb, struct virtio_mem, virtio_mem_vmcore_pfn_is_ram() local 2733 virtio_mem_vmcore_add_device_ram(struct virtio_mem * vm,struct list_head * list,uint64_t start,uint64_t end) virtio_mem_vmcore_add_device_ram() argument 2748 struct virtio_mem *vm = container_of(cb, struct virtio_mem, virtio_mem_vmcore_get_device_ram() local 2818 virtio_mem_init_kdump(struct virtio_mem * vm) virtio_mem_init_kdump() argument 2836 virtio_mem_init(struct virtio_mem * vm) virtio_mem_init() argument 2879 virtio_mem_create_resource(struct virtio_mem * vm) virtio_mem_create_resource() argument 2907 virtio_mem_delete_resource(struct virtio_mem * vm) virtio_mem_delete_resource() argument 2926 virtio_mem_has_memory_added(struct virtio_mem * vm) virtio_mem_has_memory_added() argument 2937 struct virtio_mem *vm; virtio_mem_probe() local 2984 virtio_mem_deinit_hotplug(struct virtio_mem * vm) virtio_mem_deinit_hotplug() argument 3050 virtio_mem_deinit_kdump(struct virtio_mem * vm) virtio_mem_deinit_kdump() argument 3059 struct virtio_mem *vm = vdev->priv; virtio_mem_remove() local 3076 struct virtio_mem *vm = vdev->priv; virtio_mem_config_changed() local 3088 struct virtio_mem *vm = vdev->priv; virtio_mem_freeze() local 3111 struct virtio_mem *vm = vdev->priv; virtio_mem_restore() local [all...] |
| /linux/tools/testing/selftests/kvm/lib/arm64/ |
| H A D | processor.c | 24 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) in page_align() argument 26 unsigned int shift = (vm->mmu.pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; in page_align() 27 uint64_t mask = (1UL << (vm->va_bits - shift)) - 1; in page_align() 32 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) in pgd_index() 34 unsigned int shift = 2 * (vm->page_shift - 3) + vm->page_shift; in pgd_index() 35 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pgd_index() 37 TEST_ASSERT(vm argument 29 pgd_index(struct kvm_vm * vm,vm_vaddr_t gva) pgd_index() argument 48 pmd_index(struct kvm_vm * vm,vm_vaddr_t gva) pmd_index() argument 59 pte_index(struct kvm_vm * vm,vm_vaddr_t gva) pte_index() argument 65 use_lpa2_pte_format(struct kvm_vm * vm) use_lpa2_pte_format() argument 71 addr_pte(struct kvm_vm * vm,uint64_t pa,uint64_t attrs) addr_pte() argument 89 pte_addr(struct kvm_vm * vm,uint64_t pte) pte_addr() argument 105 ptrs_per_pgd(struct kvm_vm * vm) ptrs_per_pgd() argument 111 ptrs_per_pte(struct kvm_vm * vm) ptrs_per_pte() argument 116 virt_arch_pgd_alloc(struct kvm_vm * vm) virt_arch_pgd_alloc() argument 129 _virt_pg_map(struct kvm_vm * vm,uint64_t vaddr,uint64_t paddr,uint64_t flags) _virt_pg_map() argument 182 virt_arch_pg_map(struct kvm_vm * vm,uint64_t vaddr,uint64_t paddr) virt_arch_pg_map() argument 189 virt_get_pte_hva_at_level(struct kvm_vm * vm,vm_vaddr_t gva,int level) virt_get_pte_hva_at_level() argument 233 virt_get_pte_hva(struct kvm_vm * vm,vm_vaddr_t gva) virt_get_pte_hva() argument 238 addr_arch_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_arch_gva2gpa() argument 245 pte_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent,uint64_t page,int level) pte_dump() argument 264 virt_arch_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent) virt_arch_dump() argument 281 vm_supports_el2(struct kvm_vm * vm) vm_supports_el2() argument 291 kvm_get_default_vcpu_target(struct kvm_vm * vm,struct kvm_vcpu_init * init) kvm_get_default_vcpu_target() argument 305 struct kvm_vm *vm = vcpu->vm; aarch64_vcpu_setup() local 419 __aarch64_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id,struct kvm_vcpu_init * init) __aarch64_vcpu_add() argument 438 aarch64_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id,struct kvm_vcpu_init * init,void * guest_code) aarch64_vcpu_add() argument 448 vm_arch_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id) vm_arch_vcpu_add() argument 540 vm_init_descriptor_tables(struct kvm_vm * vm) vm_init_descriptor_tables() argument 548 vm_install_sync_handler(struct kvm_vm * vm,int vector,int ec,void (* handler)(struct ex_regs *)) vm_install_sync_handler() argument 559 vm_install_exception_handler(struct kvm_vm * vm,int vector,void (* handler)(struct ex_regs *)) vm_install_exception_handler() argument 677 vm_vaddr_populate_bitmap(struct kvm_vm * vm) vm_vaddr_populate_bitmap() argument 706 kvm_arch_vm_post_create(struct kvm_vm * vm,unsigned int nr_vcpus) kvm_arch_vm_post_create() argument 717 kvm_arch_vm_finalize_vcpus(struct kvm_vm * vm) kvm_arch_vm_finalize_vcpus() argument 723 kvm_arch_vm_release(struct kvm_vm * vm) kvm_arch_vm_release() argument [all...] |
| /linux/tools/testing/selftests/kvm/s390/ |
| H A D | cmma_test.c | 97 static void create_main_memslot(struct kvm_vm *vm) in create_main_memslot() argument 101 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, 0, 0, MAIN_PAGE_COUNT, 0); in create_main_memslot() 104 vm->memslots[i] = 0; in create_main_memslot() 107 static void create_test_memslot(struct kvm_vm *vm) in create_test_memslot() argument 109 vm_userspace_mem_region_add(vm, in create_test_memslot() 111 TEST_DATA_START_GFN << vm->page_shift, in create_test_memslot() 116 vm->memslots[MEM_REGION_TEST_DATA] = TEST_DATA_MEMSLOT; in create_test_memslot() 119 static void create_memslots(struct kvm_vm *vm) in create_memslots() argument 135 create_main_memslot(vm); in create_memslots() 136 create_test_memslot(vm); in create_memslots() 139 finish_vm_setup(struct kvm_vm * vm) finish_vm_setup() argument 153 struct kvm_vm *vm; create_vm_two_memslots() local 164 enable_cmma(struct kvm_vm * vm) enable_cmma() argument 172 enable_dirty_tracking(struct kvm_vm * vm) enable_dirty_tracking() argument 178 __enable_migration_mode(struct kvm_vm * vm) __enable_migration_mode() argument 187 enable_migration_mode(struct kvm_vm * vm) enable_migration_mode() argument 194 is_migration_mode_on(struct kvm_vm * vm) is_migration_mode_on() argument 208 vm_get_cmma_bits(struct kvm_vm * vm,u64 flags,int * errno_out) vm_get_cmma_bits() argument 229 struct kvm_vm *vm = create_vm_two_memslots(); test_get_cmma_basic() local 274 struct kvm_vm *vm = vm_create_barebones(); test_migration_mode() local 376 assert_all_slots_cmma_dirty(struct kvm_vm * vm) assert_all_slots_cmma_dirty() argument 426 assert_no_pages_cmma_dirty(struct kvm_vm * vm) assert_no_pages_cmma_dirty() argument 449 struct kvm_vm *vm = create_vm_two_memslots(); test_get_initial_dirty() local 473 query_cmma_range(struct kvm_vm * vm,u64 start_gfn,u64 gfn_count,struct kvm_s390_cmma_log * res_out) query_cmma_range() argument 506 struct kvm_vm *vm = create_vm_two_memslots(); test_get_skip_holes() local 668 struct kvm_vm *vm = vm_create_barebones(); machine_has_cmma() local [all...] |
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_vm.h | 34 static inline struct xe_vm *xe_vm_get(struct xe_vm *vm) in xe_vm_get() argument 36 drm_gpuvm_get(&vm->gpuvm); in xe_vm_get() 37 return vm; in xe_vm_get() 40 static inline void xe_vm_put(struct xe_vm *vm) in xe_vm_put() argument 42 drm_gpuvm_put(&vm->gpuvm); in xe_vm_put() 45 int xe_vm_lock(struct xe_vm *vm, bool intr); 47 void xe_vm_unlock(struct xe_vm *vm); 49 static inline bool xe_vm_is_closed(struct xe_vm *vm) in xe_vm_is_closed() argument 52 return !vm->size; in xe_vm_is_closed() 55 static inline bool xe_vm_is_banned(struct xe_vm *vm) in xe_vm_is_banned() argument [all …]
|
| H A D | xe_vm.c | 44 static struct drm_gem_object *xe_vm_obj(struct xe_vm *vm) in xe_vm_obj() argument 46 return vm->gpuvm.r_obj; in xe_vm_obj() 50 * xe_vm_drm_exec_lock() - Lock the vm's resv with a drm_exec transaction 51 * @vm: The vm whose resv is to be locked. 54 * Helper to lock the vm's resv as part of a drm_exec transaction. 58 int xe_vm_drm_exec_lock(struct xe_vm *vm, struct drm_exec *exec) in xe_vm_drm_exec_lock() argument 60 return drm_exec_lock_obj(exec, xe_vm_obj(vm)); in xe_vm_drm_exec_lock() 63 static bool preempt_fences_waiting(struct xe_vm *vm) in preempt_fences_waiting() argument 67 lockdep_assert_held(&vm in preempt_fences_waiting() 89 alloc_preempt_fences(struct xe_vm * vm,struct list_head * list,unsigned int * count) alloc_preempt_fences() argument 110 wait_for_existing_preempt_fences(struct xe_vm * vm) wait_for_existing_preempt_fences() argument 142 xe_vm_is_idle(struct xe_vm * vm) xe_vm_is_idle() argument 155 arm_preempt_fences(struct xe_vm * vm,struct list_head * list) arm_preempt_fences() argument 174 add_preempt_fences(struct xe_vm * vm,struct xe_bo * bo) add_preempt_fences() argument 198 resume_and_reinstall_preempt_fences(struct xe_vm * vm,struct drm_exec * exec) resume_and_reinstall_preempt_fences() argument 214 xe_vm_add_compute_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) xe_vm_add_compute_exec_queue() argument 277 xe_vm_remove_compute_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) xe_vm_remove_compute_exec_queue() argument 305 xe_vm_kill(struct xe_vm * vm,bool unlocked) xe_vm_kill() argument 328 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); xe_gpuvm_validate() local 363 xe_vm_validate_rebind(struct xe_vm * vm,struct drm_exec * exec,unsigned int num_fences) xe_vm_validate_rebind() argument 389 xe_preempt_work_begin(struct drm_exec * exec,struct xe_vm * vm,bool * done) xe_preempt_work_begin() argument 426 vm_suspend_rebind_worker(struct xe_vm * vm) vm_suspend_rebind_worker() argument 448 xe_vm_resume_rebind_worker(struct xe_vm * vm) xe_vm_resume_rebind_worker() argument 455 struct xe_vm *vm = container_of(w, struct xe_vm, preempt.rebind_work); preempt_rebind_work_func() local 687 xe_vm_rebind(struct xe_vm * vm,bool rebind_worker) xe_vm_rebind() argument 742 xe_vma_rebind(struct xe_vm * vm,struct xe_vma * vma,u8 tile_mask) xe_vma_rebind() argument 829 xe_vm_range_rebind(struct xe_vm * vm,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_range_rebind() argument 913 xe_vm_range_unbind(struct xe_vm * vm,struct xe_svm_range * range) xe_vm_range_unbind() argument 990 xe_vma_create(struct xe_vm * vm,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 start,u64 end,struct xe_vma_mem_attr * attr,unsigned int flags) xe_vma_create() argument 1080 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_destroy_late() local 1120 struct xe_vm *vm = xe_vma_vm(vma); xe_vma_destroy() local 1160 struct xe_vm *vm = xe_vma_vm(vma); xe_vm_lock_vma() local 1191 xe_vm_find_overlapping_vma(struct xe_vm * vm,u64 start,u64 range) xe_vm_find_overlapping_vma() argument 1207 xe_vm_insert_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_insert_vma() argument 1222 xe_vm_remove_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_remove_vma() argument 1422 xe_vm_create_scratch(struct xe_device * xe,struct xe_tile * tile,struct xe_vm * vm,struct drm_exec * exec) xe_vm_create_scratch() argument 1442 xe_vm_free_scratch(struct xe_vm * vm) xe_vm_free_scratch() argument 1462 xe_vm_pt_destroy(struct xe_vm * vm) xe_vm_pt_destroy() argument 1482 struct xe_vm *vm; xe_vm_create() local 1690 xe_vm_close(struct xe_vm * vm) xe_vm_close() argument 1732 xe_vm_close_and_put(struct xe_vm * vm) xe_vm_close_and_put() argument 1843 struct xe_vm *vm = vm_destroy_work_func() local 1875 struct xe_vm *vm = container_of(gpuvm, struct xe_vm, gpuvm); xe_vm_free() local 1883 struct xe_vm *vm; xe_vm_lookup() local 1894 xe_vm_pdp4_descriptor(struct xe_vm * vm,struct xe_tile * tile) xe_vm_pdp4_descriptor() argument 1900 to_wait_exec_queue(struct xe_vm * vm,struct xe_exec_queue * q) to_wait_exec_queue() argument 1931 struct xe_vm *vm; xe_vm_create_ioctl() local 1998 struct xe_vm *vm; xe_vm_destroy_ioctl() local 2021 xe_vm_query_vmas(struct xe_vm * vm,u64 start,u64 end) xe_vm_query_vmas() argument 2033 get_mem_attrs(struct xe_vm * vm,u32 * num_vmas,u64 start,u64 end,struct drm_xe_mem_range_attr * attrs) get_mem_attrs() argument 2069 struct xe_vm *vm; xe_vm_query_vmas_attrs_ioctl() local 2140 xe_vm_find_vma_by_addr(struct xe_vm * vm,u64 page_addr) xe_vm_find_vma_by_addr() argument 2160 prep_vma_destroy(struct xe_vm * vm,struct xe_vma * vma,bool post_commit) prep_vma_destroy() argument 2220 __xe_vm_needs_clear_scratch_pages(struct xe_vm * vm,u32 bind_flags) __xe_vm_needs_clear_scratch_pages() argument 2250 vm_bind_ioctl_ops_create(struct xe_vm * vm,struct xe_vma_ops * vops,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 addr,u64 range,u32 operation,u32 flags,u32 prefetch_region,u16 pat_index) vm_bind_ioctl_ops_create() argument 2427 new_vma(struct xe_vm * vm,struct drm_gpuva_op_map * op,struct xe_vma_mem_attr * attr,unsigned int flags) new_vma() argument 2529 xe_vma_op_commit(struct xe_vm * vm,struct xe_vma_op * op) xe_vma_op_commit() argument 2615 vm_bind_ioctl_ops_parse(struct xe_vm * vm,struct drm_gpuva_ops * ops,struct xe_vma_ops * vops) vm_bind_ioctl_ops_parse() argument 2796 xe_vma_op_unwind(struct xe_vm * vm,struct xe_vma_op * op,bool post_commit,bool prev_post_commit,bool next_post_commit) xe_vma_op_unwind() argument 2851 vm_bind_ioctl_ops_unwind(struct xe_vm * vm,struct drm_gpuva_ops ** ops,int num_ops_list) vm_bind_ioctl_ops_unwind() argument 2879 struct xe_vm *vm = xe_vma_vm(vma); vma_lock_and_validate() local 2909 prefetch_ranges(struct xe_vm * vm,struct xe_vma_op * op) prefetch_ranges() argument 2964 op_lock_and_prep(struct drm_exec * exec,struct xe_vm * vm,struct xe_vma_ops * vops,struct xe_vma_op * op) op_lock_and_prep() argument 3037 vm_bind_ioctl_ops_prefetch_ranges(struct xe_vm * vm,struct xe_vma_ops * vops) vm_bind_ioctl_ops_prefetch_ranges() argument 3057 vm_bind_ioctl_ops_lock_and_prep(struct drm_exec * exec,struct xe_vm * vm,struct xe_vma_ops * vops) vm_bind_ioctl_ops_lock_and_prep() argument 3116 vm_ops_setup_tile_args(struct xe_vm * vm,struct xe_vma_ops * vops) vm_ops_setup_tile_args() argument 3142 ops_execute(struct xe_vm * vm,struct xe_vma_ops * vops) ops_execute() argument 3250 op_add_ufence(struct xe_vm * vm,struct xe_vma_op * op,struct xe_user_fence * ufence) op_add_ufence() argument 3274 vm_bind_ioctl_ops_fini(struct xe_vm * vm,struct xe_vma_ops * vops,struct dma_fence * fence) vm_bind_ioctl_ops_fini() argument 3300 vm_bind_ioctl_ops_execute(struct xe_vm * vm,struct xe_vma_ops * vops) vm_bind_ioctl_ops_execute() argument 3355 vm_bind_ioctl_check_args(struct xe_device * xe,struct xe_vm * vm,struct drm_xe_vm_bind * args,struct drm_xe_vm_bind_op ** bind_ops) vm_bind_ioctl_check_args() argument 3493 vm_bind_ioctl_signal_fences(struct xe_vm * vm,struct xe_exec_queue * q,struct xe_sync_entry * syncs,int num_syncs) vm_bind_ioctl_signal_fences() argument 3516 xe_vma_ops_init(struct xe_vma_ops * vops,struct xe_vm * vm,struct xe_exec_queue * q,struct xe_sync_entry * syncs,u32 num_syncs) xe_vma_ops_init() argument 3605 struct xe_vm *vm; xe_vm_bind_ioctl() local 3847 xe_vm_bind_kernel_bo(struct xe_vm * vm,struct xe_bo * bo,struct xe_exec_queue * q,u64 addr,enum xe_cache_level cache_lvl) xe_vm_bind_kernel_bo() argument 3917 xe_vm_lock(struct xe_vm * vm,bool intr) xe_vm_lock() argument 3935 xe_vm_unlock(struct xe_vm * vm) xe_vm_unlock() argument 3952 xe_vm_range_tilemask_tlb_inval(struct xe_vm * vm,u64 start,u64 end,u8 tile_mask) xe_vm_range_tilemask_tlb_inval() argument 4013 struct xe_vm *vm = xe_vma_vm(vma); xe_vm_invalidate_vma() local 4063 xe_vm_validate_protected(struct xe_vm * vm) xe_vm_validate_protected() argument 4111 xe_vm_snapshot_capture(struct xe_vm * vm) xe_vm_snapshot_capture() argument 4355 xe_vm_alloc_vma(struct xe_vm * vm,struct drm_gpuvm_map_req * map_req,bool is_madvise) xe_vm_alloc_vma() argument 4495 xe_vm_alloc_madvise_vma(struct xe_vm * vm,uint64_t start,uint64_t range) xe_vm_alloc_madvise_vma() argument 4530 xe_vm_find_cpu_addr_mirror_vma_range(struct xe_vm * vm,u64 * start,u64 * end) xe_vm_find_cpu_addr_mirror_vma_range() argument 4559 xe_vm_alloc_cpu_addr_mirror_vma(struct xe_vm * vm,uint64_t start,uint64_t range) xe_vm_alloc_cpu_addr_mirror_vma() argument [all...] |
| H A D | xe_exec.c | 101 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn() local 105 xe_vm_set_validation_exec(vm, &vm_exec->exec); in xe_exec_fn() 106 ret = xe_vm_validate_rebind(vm, &vm_exec->exec, 1); in xe_exec_fn() 107 xe_vm_set_validation_exec(vm, NULL); in xe_exec_fn() 126 struct xe_vm *vm; in xe_exec_ioctl() local 172 vm = q->vm; in xe_exec_ioctl() 178 (xe_vm_in_lr_mode(vm) ? in xe_exec_ioctl() 216 if (!xe_vm_in_lr_mode(vm) in xe_exec_ioctl() [all...] |
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_vm.c | 120 * @vm: pointer to the amdgpu_vm structure to set the fence sequence on 122 struct amdgpu_vm *vm; member 132 * @vm: the VM which schould be tested 136 static void amdgpu_vm_assert_locked(struct amdgpu_vm *vm) in amdgpu_vm_assert_locked() argument 138 dma_resv_assert_held(vm->root.bo->tbo.base.resv); in amdgpu_vm_assert_locked() 151 struct amdgpu_vm *vm = vm_bo->vm; in amdgpu_vm_bo_evicted() local 155 amdgpu_vm_assert_locked(vm); in amdgpu_vm_bo_evicted() 156 spin_lock(&vm_bo->vm->status_lock); in amdgpu_vm_bo_evicted() 158 list_move(&vm_bo->vm_status, &vm in amdgpu_vm_bo_evicted() 270 amdgpu_vm_bo_reset_state_machine(struct amdgpu_vm * vm) amdgpu_vm_bo_reset_state_machine() argument 303 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_shared() local 353 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_stats_locked() local 392 struct amdgpu_vm *vm = base->vm; amdgpu_vm_update_stats() local 410 amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base * base,struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_base_init() argument 459 amdgpu_vm_lock_pd(struct amdgpu_vm * vm,struct drm_exec * exec,unsigned int num_fences) amdgpu_vm_lock_pd() argument 475 amdgpu_vm_lock_done_list(struct amdgpu_vm * vm,struct drm_exec * exec,unsigned int num_fences) amdgpu_vm_lock_done_list() argument 517 amdgpu_vm_move_to_lru_tail(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_move_to_lru_tail() argument 526 amdgpu_vm_init_entities(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_init_entities() argument 546 amdgpu_vm_fini_entities(struct amdgpu_vm * vm) amdgpu_vm_fini_entities() argument 561 amdgpu_vm_generation(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_generation() argument 592 amdgpu_vm_validate(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket,int (* validate)(void * p,struct amdgpu_bo * bo),void * param) amdgpu_vm_validate() argument 668 amdgpu_vm_ready(struct amdgpu_vm * vm) amdgpu_vm_ready() argument 913 amdgpu_vm_bo_find(struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_find() argument 967 amdgpu_vm_update_pdes(struct amdgpu_device * adev,struct amdgpu_vm * vm,bool immediate) amdgpu_vm_update_pdes() argument 1055 struct amdgpu_vm *vm = params->vm; amdgpu_vm_tlb_flush() local 1105 amdgpu_vm_update_range(struct amdgpu_device * adev,struct amdgpu_vm * vm,bool immediate,bool unlocked,bool flush_tlb,bool allow_override,struct amdgpu_sync * sync,uint64_t start,uint64_t last,uint64_t flags,uint64_t offset,uint64_t vram_base,struct ttm_resource * res,dma_addr_t * pages_addr,struct dma_fence ** fence) amdgpu_vm_update_range() argument 1239 amdgpu_vm_get_memory(struct amdgpu_vm * vm,struct amdgpu_mem_stats stats[__AMDGPU_PL_NUM]) amdgpu_vm_get_memory() argument 1263 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_update() local 1500 amdgpu_vm_free_mapping(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_bo_va_mapping * mapping,struct dma_fence * fence) amdgpu_vm_free_mapping() argument 1517 amdgpu_vm_prt_fini(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_prt_fini() argument 1546 amdgpu_vm_clear_freed(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct dma_fence ** fence) amdgpu_vm_clear_freed() argument 1608 amdgpu_vm_handle_moved(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket) amdgpu_vm_handle_moved() argument 1686 amdgpu_vm_flush_compute_tlb(struct amdgpu_device * adev,struct amdgpu_vm * vm,uint32_t flush_type,uint32_t xcc_mask) amdgpu_vm_flush_compute_tlb() argument 1733 amdgpu_vm_bo_add(struct amdgpu_device * adev,struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_bo_add() argument 1778 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_insert_map() local 1849 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_map() local 1958 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_unmap() local 2020 amdgpu_vm_bo_clear_mappings(struct amdgpu_device * adev,struct amdgpu_vm * vm,uint64_t saddr,uint64_t size) amdgpu_vm_bo_clear_mappings() argument 2138 amdgpu_vm_bo_lookup_mapping(struct amdgpu_vm * vm,uint64_t addr) amdgpu_vm_bo_lookup_mapping() argument 2152 amdgpu_vm_bo_trace_cs(struct amdgpu_vm * vm,struct ww_acquire_ctx * ticket) amdgpu_vm_bo_trace_cs() argument 2189 struct amdgpu_vm *vm = bo_va->base.vm; amdgpu_vm_bo_del() local 2283 struct amdgpu_vm *vm = bo_base->vm; amdgpu_vm_bo_invalidate() local 2318 struct amdgpu_vm *vm = bo_base->vm; amdgpu_vm_bo_move() local 2455 amdgpu_vm_wait_idle(struct amdgpu_vm * vm,long timeout) amdgpu_vm_wait_idle() argument 2474 struct amdgpu_vm *vm; amdgpu_vm_get_vm_from_pasid() local 2506 amdgpu_vm_get_task_info_vm(struct amdgpu_vm * vm) amdgpu_vm_get_task_info_vm() argument 2534 amdgpu_vm_create_task_info(struct amdgpu_vm * vm) amdgpu_vm_create_task_info() argument 2549 amdgpu_vm_set_task_info(struct amdgpu_vm * vm) amdgpu_vm_set_task_info() argument 2577 amdgpu_vm_init(struct amdgpu_device * adev,struct amdgpu_vm * vm,int32_t xcp_id,uint32_t pasid) amdgpu_vm_init() argument 2706 amdgpu_vm_make_compute(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_make_compute() argument 2748 amdgpu_vm_stats_is_zero(struct amdgpu_vm * vm) amdgpu_vm_stats_is_zero() argument 2767 amdgpu_vm_fini(struct amdgpu_device * adev,struct amdgpu_vm * vm) amdgpu_vm_fini() argument 2913 struct amdgpu_vm *vm = &fpriv->vm; amdgpu_vm_ioctl() local 2955 struct amdgpu_vm *vm; amdgpu_vm_handle_fault() local 3045 amdgpu_debugfs_vm_bo_info(struct amdgpu_vm * vm,struct seq_file * m) amdgpu_debugfs_vm_bo_info() argument 3150 struct amdgpu_vm *vm; amdgpu_vm_update_fault_cache() local 3200 amdgpu_vm_is_bo_always_valid(struct amdgpu_vm * vm,struct amdgpu_bo * bo) amdgpu_vm_is_bo_always_valid() argument [all...] |
| /linux/tools/testing/selftests/kvm/lib/loongarch/ |
| H A D | processor.c | 17 static uint64_t virt_pte_index(struct kvm_vm *vm, vm_vaddr_t gva, int level) in virt_pte_index() argument 22 shift = level * (vm->page_shift - 3) + vm->page_shift; in virt_pte_index() 23 mask = (1UL << (vm->page_shift - 3)) - 1; in virt_pte_index() 27 static uint64_t pte_addr(struct kvm_vm *vm, uint64_t entry) in pte_addr() argument 29 return entry & ~((0x1UL << vm->page_shift) - 1); in pte_addr() 32 static uint64_t ptrs_per_pte(struct kvm_vm *vm) in ptrs_per_pte() argument 34 return 1 << (vm->page_shift - 3); in ptrs_per_pte() 37 static void virt_set_pgtable(struct kvm_vm *vm, vm_paddr_t table, vm_paddr_t child) in virt_set_pgtable() argument 42 ptep = addr_gpa2hva(vm, tabl in virt_set_pgtable() 48 virt_arch_pgd_alloc(struct kvm_vm * vm) virt_arch_pgd_alloc() argument 74 virt_populate_pte(struct kvm_vm * vm,vm_vaddr_t gva,int alloc) virt_populate_pte() argument 108 addr_arch_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_arch_gva2gpa() argument 118 virt_arch_pg_map(struct kvm_vm * vm,uint64_t vaddr,uint64_t paddr) virt_arch_pg_map() argument 142 pte_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent,uint64_t page,int level) pte_dump() argument 160 virt_arch_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent) virt_arch_dump() argument 204 vm_init_descriptor_tables(struct kvm_vm * vm) vm_init_descriptor_tables() argument 217 vm_install_exception_handler(struct kvm_vm * vm,int vector,handler_fn handler) vm_install_exception_handler() argument 274 struct kvm_vm *vm = vcpu->vm; loongarch_vcpu_setup() local 359 vm_arch_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id) vm_arch_vcpu_add() argument [all...] |
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_mmu.c | 47 /** @vm: VM bound to this slot. NULL is no VM is bound. */ 48 struct panthor_vm *vm; member 91 /** @vm: VMs management fields */ 93 /** @vm.lock: Lock protecting access to list. */ 96 /** @vm.list: List containing all VMs. */ 99 /** @vm.reset_in_progress: True if a reset is in progress. */ 102 /** @vm.wq: Workqueue used for the VM_BIND queues. */ 104 } vm; member 338 /** @node: Used to insert the VM in the panthor_mmu::vm::list. */ 402 /** @vm 403 struct panthor_vm *vm; global() member 435 struct panthor_vm *vm = cookie; alloc_pt() local 487 struct panthor_vm *vm = cookie; free_pt() local 578 struct panthor_vm *vm = ptdev->mmu->as.slots[as_nr].vm; panthor_mmu_as_disable() local 626 panthor_vm_has_unhandled_faults(struct panthor_vm * vm) panthor_vm_has_unhandled_faults() argument 637 panthor_vm_is_unusable(struct panthor_vm * vm) panthor_vm_is_unusable() argument 642 panthor_vm_release_as_locked(struct panthor_vm * vm) panthor_vm_release_as_locked() argument 666 panthor_vm_active(struct panthor_vm * vm) panthor_vm_active() argument 790 panthor_vm_idle(struct panthor_vm * vm) panthor_vm_idle() argument 804 panthor_vm_page_size(struct panthor_vm * vm) panthor_vm_page_size() argument 812 panthor_vm_stop(struct panthor_vm * vm) panthor_vm_stop() argument 817 panthor_vm_start(struct panthor_vm * vm) panthor_vm_start() argument 828 panthor_vm_as(struct panthor_vm * vm) panthor_vm_as() argument 854 panthor_vm_declare_unusable(struct panthor_vm * vm) panthor_vm_declare_unusable() argument 871 panthor_vm_unmap_pages(struct panthor_vm * vm,u64 iova,u64 size) panthor_vm_unmap_pages() argument 924 panthor_vm_map_pages(struct panthor_vm * vm,u64 iova,int prot,struct sg_table * sgt,u64 offset,u64 size) panthor_vm_map_pages() argument 1031 panthor_vm_alloc_va(struct panthor_vm * vm,u64 va,u64 size,struct drm_mm_node * va_node) panthor_vm_alloc_va() argument 1065 panthor_vm_free_va(struct panthor_vm * vm,struct drm_mm_node * va_node) panthor_vm_free_va() argument 1082 panthor_vm_cleanup_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm) panthor_vm_cleanup_op_ctx() argument 1180 panthor_vm_prepare_map_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm,struct panthor_gem_object * bo,u64 offset,u64 size,u64 va,u32 flags) panthor_vm_prepare_map_op_ctx() argument 1288 panthor_vm_prepare_unmap_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm,u64 va,u64 size) panthor_vm_prepare_unmap_op_ctx() argument 1339 panthor_vm_prepare_sync_only_op_ctx(struct panthor_vm_op_ctx * op_ctx,struct panthor_vm * vm) panthor_vm_prepare_sync_only_op_ctx() argument 1360 panthor_vm_get_bo_for_va(struct panthor_vm * vm,u64 va,u64 * bo_offset) panthor_vm_get_bo_for_va() argument 1465 struct panthor_vm *vm; panthor_vm_pool_create_vm() local 1490 panthor_vm_destroy(struct panthor_vm * vm) panthor_vm_destroy() argument 1529 struct panthor_vm *vm; panthor_vm_pool_destroy_vm() local 1548 struct panthor_vm *vm; panthor_vm_pool_get_vm() local 1568 struct panthor_vm *vm; panthor_vm_pool_destroy() local 1629 panthor_vm_lock_region(struct panthor_vm * vm,u64 start,u64 size) panthor_vm_lock_region() argument 1661 panthor_vm_unlock_region(struct panthor_vm * vm) panthor_vm_unlock_region() argument 1778 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_suspend() local 1822 struct panthor_vm *vm; panthor_mmu_pre_reset() local 1842 struct panthor_vm *vm; panthor_mmu_post_reset() local 1853 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_post_reset() local 1874 struct panthor_vm *vm = container_of(gpuvm, struct panthor_vm, base); panthor_vm_free() local 1924 panthor_vm_put(struct panthor_vm * vm) panthor_vm_put() argument 1935 panthor_vm_get(struct panthor_vm * vm) panthor_vm_get() argument 1957 panthor_vm_get_heap_pool(struct panthor_vm * vm,bool create) panthor_vm_get_heap_pool() argument 1991 struct panthor_vm *vm; panthor_vm_heaps_sizes() local 2049 panthor_vma_link(struct panthor_vm * vm,struct panthor_vma * vma,struct drm_gpuvm_bo * vm_bo) panthor_vma_link() argument 2079 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_map() local 2152 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_remap() local 2229 struct panthor_vm *vm = priv; panthor_gpuva_sm_step_unmap() local 2252 panthor_vm_resv(struct panthor_vm * vm) panthor_vm_resv() argument 2257 panthor_vm_root_gem(struct panthor_vm * vm) panthor_vm_root_gem() argument 2266 panthor_vm_exec_op(struct panthor_vm * vm,struct panthor_vm_op_ctx * op,bool flag_vm_unusable_on_failure) panthor_vm_exec_op() argument 2425 struct panthor_vm *vm; panthor_vm_create() local 2527 panthor_vm_bind_prepare_op_ctx(struct drm_file * file,struct panthor_vm * vm,const struct drm_panthor_vm_bind_op * op,struct panthor_vm_op_ctx * op_ctx) panthor_vm_bind_prepare_op_ctx() argument 2599 panthor_vm_bind_job_create(struct drm_file * file,struct panthor_vm * vm,const struct drm_panthor_vm_bind_op * op) panthor_vm_bind_job_create() argument 2685 panthor_vm_update_resvs(struct panthor_vm * vm,struct drm_exec * exec,struct dma_fence * fence,enum dma_resv_usage private_usage,enum dma_resv_usage extobj_usage) panthor_vm_update_resvs() argument 2702 panthor_vm_bind_exec_sync_op(struct drm_file * file,struct panthor_vm * vm,struct drm_panthor_vm_bind_op * op) panthor_vm_bind_exec_sync_op() argument 2740 panthor_vm_map_bo_range(struct panthor_vm * vm,struct panthor_gem_object * bo,u64 offset,u64 size,u64 va,u32 flags) panthor_vm_map_bo_range() argument 2767 panthor_vm_unmap_range(struct panthor_vm * vm,u64 va,u64 size) panthor_vm_unmap_range() argument 2795 panthor_vm_prepare_mapped_bos_resvs(struct drm_exec * exec,struct panthor_vm * vm,u32 slot_count) panthor_vm_prepare_mapped_bos_resvs() argument 2822 struct panthor_vm *vm = ptdev->mmu->as.slots[i].vm; panthor_mmu_unplug() local 2893 show_vm_gpuvas(struct panthor_vm * vm,struct seq_file * m) show_vm_gpuvas() argument 2910 struct panthor_vm *vm; show_each_vm() local [all...] |
| /linux/drivers/gpu/drm/lima/ |
| H A D | lima_vm.c | 18 struct lima_vm *vm; member 35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) in lima_vm_unmap_range() argument 43 vm->bts[pbe].cpu[bte] = 0; in lima_vm_unmap_range() 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument 52 if (!vm->bts[pbe].cpu) { in lima_vm_map_page() 57 vm->bts[pbe].cpu = dma_alloc_wc( in lima_vm_map_page() 58 vm->dev->dev, LIMA_PAGE_SIZE << LIMA_VM_NUM_PT_PER_BT_SHIFT, in lima_vm_map_page() 59 &vm->bts[pbe].dma, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO); in lima_vm_map_page() 60 if (!vm->bts[pbe].cpu) in lima_vm_map_page() 63 pts = vm in lima_vm_map_page() 77 lima_vm_bo_find(struct lima_vm * vm,struct lima_bo * bo) lima_vm_bo_find() argument 91 lima_vm_bo_add(struct lima_vm * vm,struct lima_bo * bo,bool create) lima_vm_bo_add() argument 155 lima_vm_bo_del(struct lima_vm * vm,struct lima_bo * bo) lima_vm_bo_del() argument 185 lima_vm_get_va(struct lima_vm * vm,struct lima_bo * bo) lima_vm_get_va() argument 202 struct lima_vm *vm; lima_vm_create() local 237 struct lima_vm *vm = container_of(kref, struct lima_vm, refcount); lima_vm_release() local 254 lima_vm_print(struct lima_vm * vm) lima_vm_print() argument 283 lima_vm_map_bo(struct lima_vm * vm,struct lima_bo * bo,int pageoff) lima_vm_map_bo() argument [all...] |
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | kvm_util.c | 168 void vm_enable_dirty_ring(struct kvm_vm *vm, uint32_t ring_size) in vm_enable_dirty_ring() argument 170 if (vm_check_cap(vm, KVM_CAP_DIRTY_LOG_RING_ACQ_REL)) in vm_enable_dirty_ring() 171 vm_enable_cap(vm, KVM_CAP_DIRTY_LOG_RING_ACQ_REL, ring_size); in vm_enable_dirty_ring() 173 vm_enable_cap(vm, KVM_CAP_DIRTY_LOG_RING, ring_size); in vm_enable_dirty_ring() 174 vm->dirty_ring_size = ring_size; in vm_enable_dirty_ring() 177 static void vm_open(struct kvm_vm *vm) in vm_open() argument 179 vm->kvm_fd = _open_kvm_dev_path_or_exit(O_RDWR); in vm_open() 183 vm->fd = __kvm_ioctl(vm->kvm_fd, KVM_CREATE_VM, (void *)vm in vm_open() 252 vm_vaddr_populate_bitmap(struct kvm_vm * vm) vm_vaddr_populate_bitmap() argument 263 struct kvm_vm *vm; ____vm_create() local 452 struct kvm_vm *vm; __vm_create() local 520 struct kvm_vm *vm; __vm_create_with_vcpus() local 540 struct kvm_vm *vm; __vm_create_shape_with_one_vcpu() local 583 vm_arch_vcpu_recreate(struct kvm_vm * vm,uint32_t vcpu_id) vm_arch_vcpu_recreate() argument 589 vm_recreate_with_one_vcpu(struct kvm_vm * vm) vm_recreate_with_one_vcpu() argument 685 userspace_mem_region_find(struct kvm_vm * vm,uint64_t start,uint64_t end) userspace_mem_region_find() argument 738 vm_vcpu_rm(struct kvm_vm * vm,struct kvm_vcpu * vcpu) vm_vcpu_rm() argument 772 __vm_mem_region_delete(struct kvm_vm * vm,struct userspace_mem_region * region) __vm_mem_region_delete() argument 888 __vm_set_user_memory_region(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva) __vm_set_user_memory_region() argument 902 vm_set_user_memory_region(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva) vm_set_user_memory_region() argument 915 __vm_set_user_memory_region2(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva,uint32_t guest_memfd,uint64_t guest_memfd_offset) __vm_set_user_memory_region2() argument 934 vm_set_user_memory_region2(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva,uint32_t guest_memfd,uint64_t guest_memfd_offset) vm_set_user_memory_region2() argument 947 vm_mem_add(struct kvm_vm * vm,enum vm_mem_backing_src_type src_type,uint64_t gpa,uint32_t slot,uint64_t npages,uint32_t flags,int guest_memfd,uint64_t guest_memfd_offset) vm_mem_add() argument 1116 vm_userspace_mem_region_add(struct kvm_vm * vm,enum vm_mem_backing_src_type src_type,uint64_t gpa,uint32_t slot,uint64_t npages,uint32_t flags) vm_userspace_mem_region_add() argument 1140 memslot2region(struct kvm_vm * vm,uint32_t memslot) memslot2region() argument 1171 vm_mem_region_set_flags(struct kvm_vm * vm,uint32_t slot,uint32_t flags) vm_mem_region_set_flags() argument 1187 vm_mem_region_reload(struct kvm_vm * vm,uint32_t slot) vm_mem_region_reload() argument 1211 vm_mem_region_move(struct kvm_vm * vm,uint32_t slot,uint64_t new_gpa) vm_mem_region_move() argument 1240 vm_mem_region_delete(struct kvm_vm * vm,uint32_t slot) vm_mem_region_delete() argument 1250 vm_guest_mem_fallocate(struct kvm_vm * vm,uint64_t base,uint64_t size,bool punch_hole) vm_guest_mem_fallocate() argument 1294 vcpu_exists(struct kvm_vm * vm,uint32_t vcpu_id) vcpu_exists() argument 1310 __vm_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id) __vm_vcpu_add() argument 1363 vm_vaddr_unused_gap(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min) vm_vaddr_unused_gap() argument 1429 ____vm_vaddr_alloc(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min,enum kvm_mem_region_type type,bool protected) ____vm_vaddr_alloc() argument 1457 __vm_vaddr_alloc(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min,enum kvm_mem_region_type type) __vm_vaddr_alloc() argument 1464 vm_vaddr_alloc_shared(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min,enum kvm_mem_region_type type) vm_vaddr_alloc_shared() argument 1490 vm_vaddr_alloc(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min) vm_vaddr_alloc() argument 1509 vm_vaddr_alloc_pages(struct kvm_vm * vm,int nr_pages) vm_vaddr_alloc_pages() argument 1514 __vm_vaddr_alloc_page(struct kvm_vm * vm,enum kvm_mem_region_type type) __vm_vaddr_alloc_page() argument 1533 vm_vaddr_alloc_page(struct kvm_vm * vm) vm_vaddr_alloc_page() argument 1554 virt_map(struct kvm_vm * vm,uint64_t vaddr,uint64_t paddr,unsigned int npages) virt_map() argument 1588 addr_gpa2hva(struct kvm_vm * vm,vm_paddr_t gpa) addr_gpa2hva() argument 1621 addr_hva2gpa(struct kvm_vm * vm,void * hva) addr_hva2gpa() argument 1664 addr_gpa2alias(struct kvm_vm * vm,vm_paddr_t gpa) addr_gpa2alias() argument 1681 vm_create_irqchip(struct kvm_vm * vm) vm_create_irqchip() argument 1799 __kvm_test_create_device(struct kvm_vm * vm,uint64_t type) __kvm_test_create_device() argument 1809 __kvm_create_device(struct kvm_vm * vm,uint64_t type) __kvm_create_device() argument 1851 _kvm_irq_line(struct kvm_vm * vm,uint32_t irq,int level) _kvm_irq_line() argument 1861 kvm_irq_line(struct kvm_vm * vm,uint32_t irq,int level) kvm_irq_line() argument 1899 _kvm_gsi_routing_write(struct kvm_vm * vm,struct kvm_irq_routing * routing) _kvm_gsi_routing_write() argument 1910 kvm_gsi_routing_write(struct kvm_vm * vm,struct kvm_irq_routing * routing) kvm_gsi_routing_write() argument 1933 vm_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent) vm_dump() argument 2068 __vm_phy_pages_alloc(struct kvm_vm * vm,size_t num,vm_paddr_t paddr_min,uint32_t memslot,bool protected) __vm_phy_pages_alloc() argument 2114 vm_phy_page_alloc(struct kvm_vm * vm,vm_paddr_t paddr_min,uint32_t memslot) vm_phy_page_alloc() argument 2120 vm_alloc_page_table(struct kvm_vm * vm) vm_alloc_page_table() argument 2138 addr_gva2hva(struct kvm_vm * vm,vm_vaddr_t gva) addr_gva2hva() argument 2143 vm_compute_max_gfn(struct kvm_vm * vm) vm_compute_max_gfn() argument 2283 kvm_arch_vm_post_create(struct kvm_vm * vm,unsigned int nr_vcpus) kvm_arch_vm_post_create() argument 2287 kvm_arch_vm_finalize_vcpus(struct kvm_vm * vm) kvm_arch_vm_finalize_vcpus() argument 2291 kvm_arch_vm_release(struct kvm_vm * vm) kvm_arch_vm_release() argument 2334 vm_is_gpa_protected(struct kvm_vm * vm,vm_paddr_t paddr) vm_is_gpa_protected() argument [all...] |
| /linux/tools/testing/selftests/kvm/lib/s390/ |
| H A D | processor.c | 13 void virt_arch_pgd_alloc(struct kvm_vm *vm) in virt_arch_pgd_alloc() argument 17 TEST_ASSERT(vm->page_size == PAGE_SIZE, "Unsupported page size: 0x%x", in virt_arch_pgd_alloc() 18 vm->page_size); in virt_arch_pgd_alloc() 20 if (vm->mmu.pgd_created) in virt_arch_pgd_alloc() 23 paddr = vm_phy_pages_alloc(vm, PAGES_PER_REGION, in virt_arch_pgd_alloc() 25 vm->memslots[MEM_REGION_PT]); in virt_arch_pgd_alloc() 26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_arch_pgd_alloc() 28 vm->mmu.pgd = paddr; in virt_arch_pgd_alloc() 29 vm in virt_arch_pgd_alloc() 37 virt_alloc_region(struct kvm_vm * vm,int ri) virt_alloc_region() argument 50 virt_arch_pg_map(struct kvm_vm * vm,uint64_t gva,uint64_t gpa) virt_arch_pg_map() argument 89 addr_arch_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_arch_gva2gpa() argument 114 virt_dump_ptes(FILE * stream,struct kvm_vm * vm,uint8_t indent,uint64_t ptea_start) virt_dump_ptes() argument 128 virt_dump_region(FILE * stream,struct kvm_vm * vm,uint8_t indent,uint64_t reg_tab_addr) virt_dump_region() argument 150 virt_arch_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent) virt_arch_dump() argument 163 vm_arch_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id) vm_arch_vcpu_add() argument [all...] |
| /linux/drivers/virt/acrn/ |
| H A D | vm.c | 25 struct acrn_vm *acrn_vm_create(struct acrn_vm *vm, in acrn_vm_create() argument 37 mutex_init(&vm->regions_mapping_lock); in acrn_vm_create() 38 INIT_LIST_HEAD(&vm->ioreq_clients); in acrn_vm_create() 39 spin_lock_init(&vm->ioreq_clients_lock); in acrn_vm_create() 40 vm->vmid = vm_param->vmid; in acrn_vm_create() 41 vm->vcpu_num = vm_param->vcpu_num; in acrn_vm_create() 43 if (acrn_ioreq_init(vm, vm_param->ioreq_buf) < 0) { in acrn_vm_create() 45 vm->vmid = ACRN_INVALID_VMID; in acrn_vm_create() 50 list_add(&vm->list, &acrn_vm_list); in acrn_vm_create() 53 acrn_ioeventfd_init(vm); in acrn_vm_create() [all …]
|
| H A D | irqfd.c | 21 * @vm: Associated VM pointer 30 struct acrn_vm *vm; member 41 struct acrn_vm *vm = irqfd->vm; in acrn_irqfd_inject() local 43 acrn_msi_inject(vm, irqfd->msi.msi_addr, in acrn_irqfd_inject() 51 lockdep_assert_held(&irqfd->vm->irqfds_lock); in hsm_irqfd_shutdown() 63 struct acrn_vm *vm; in hsm_irqfd_shutdown_work() local 66 vm = irqfd->vm; in hsm_irqfd_shutdown_work() 67 mutex_lock(&vm in hsm_irqfd_shutdown_work() 79 struct acrn_vm *vm; hsm_irqfd_wakeup() local 108 acrn_irqfd_assign(struct acrn_vm * vm,struct acrn_irqfd * args) acrn_irqfd_assign() argument 170 acrn_irqfd_deassign(struct acrn_vm * vm,struct acrn_irqfd * args) acrn_irqfd_deassign() argument 193 acrn_irqfd_config(struct acrn_vm * vm,struct acrn_irqfd * args) acrn_irqfd_config() argument 205 acrn_irqfd_init(struct acrn_vm * vm) acrn_irqfd_init() argument 217 acrn_irqfd_deinit(struct acrn_vm * vm) acrn_irqfd_deinit() argument [all...] |
| H A D | ioreq.c | 39 static int ioreq_complete_request(struct acrn_vm *vm, u16 vcpu, in ioreq_complete_request() argument 64 ret = hcall_notify_req_finish(vm->vmid, vcpu); in ioreq_complete_request() 79 if (vcpu >= client->vm->vcpu_num) in acrn_ioreq_complete_request() 84 acrn_req = (struct acrn_io_request *)client->vm->ioreq_buf; in acrn_ioreq_complete_request() 88 ret = ioreq_complete_request(client->vm, vcpu, acrn_req); in acrn_ioreq_complete_request() 93 int acrn_ioreq_request_default_complete(struct acrn_vm *vm, u16 vcpu) in acrn_ioreq_request_default_complete() argument 97 spin_lock_bh(&vm->ioreq_clients_lock); in acrn_ioreq_request_default_complete() 98 if (vm->default_client) in acrn_ioreq_request_default_complete() 99 ret = acrn_ioreq_complete_request(vm->default_client, in acrn_ioreq_request_default_complete() 101 spin_unlock_bh(&vm in acrn_ioreq_request_default_complete() 212 acrn_ioreq_request_clear(struct acrn_vm * vm) acrn_ioreq_request_clear() argument 312 handle_cf8cfc(struct acrn_vm * vm,struct acrn_io_request * req,u16 vcpu) handle_cf8cfc() argument 381 find_ioreq_client(struct acrn_vm * vm,struct acrn_io_request * req) find_ioreq_client() argument 415 acrn_ioreq_client_create(struct acrn_vm * vm,ioreq_handler_t handler,void * priv,bool is_default,const char * name) acrn_ioreq_client_create() argument 468 struct acrn_vm *vm = client->vm; acrn_ioreq_client_destroy() local 497 acrn_ioreq_dispatch(struct acrn_vm * vm) acrn_ioreq_dispatch() argument 546 struct acrn_vm *vm; ioreq_dispatcher() local 596 acrn_ioreq_init(struct acrn_vm * vm,u64 buf_vma) acrn_ioreq_init() argument 636 acrn_ioreq_deinit(struct acrn_vm * vm) acrn_ioreq_deinit() argument [all...] |
| H A D | ioeventfd.c | 43 static void acrn_ioeventfd_shutdown(struct acrn_vm *vm, struct hsm_ioeventfd *p) in acrn_ioeventfd_shutdown() argument 45 lockdep_assert_held(&vm->ioeventfds_lock); in acrn_ioeventfd_shutdown() 52 static bool hsm_ioeventfd_is_conflict(struct acrn_vm *vm, in hsm_ioeventfd_is_conflict() argument 57 lockdep_assert_held(&vm->ioeventfds_lock); in hsm_ioeventfd_is_conflict() 60 list_for_each_entry(p, &vm->ioeventfds, list) in hsm_ioeventfd_is_conflict() 76 static int acrn_ioeventfd_assign(struct acrn_vm *vm, in acrn_ioeventfd_assign() argument 121 mutex_lock(&vm->ioeventfds_lock); in acrn_ioeventfd_assign() 123 if (hsm_ioeventfd_is_conflict(vm, p)) { in acrn_ioeventfd_assign() 129 ret = acrn_ioreq_range_add(vm->ioeventfd_client, p->type, in acrn_ioeventfd_assign() 134 list_add_tail(&p->list, &vm in acrn_ioeventfd_assign() 147 acrn_ioeventfd_deassign(struct acrn_vm * vm,struct acrn_ioeventfd * args) acrn_ioeventfd_deassign() argument 173 hsm_ioeventfd_match(struct acrn_vm * vm,u64 addr,u64 data,int len,int type) hsm_ioeventfd_match() argument 232 acrn_ioeventfd_config(struct acrn_vm * vm,struct acrn_ioeventfd * args) acrn_ioeventfd_config() argument 244 acrn_ioeventfd_init(struct acrn_vm * vm) acrn_ioeventfd_init() argument 263 acrn_ioeventfd_deinit(struct acrn_vm * vm) acrn_ioeventfd_deinit() argument [all...] |
| /linux/drivers/gpu/drm/i915/selftests/ |
| H A D | mock_gtt.c | 27 static void mock_insert_page(struct i915_address_space *vm, in mock_insert_page() argument 35 static void mock_insert_entries(struct i915_address_space *vm, in mock_insert_entries() argument 41 static void mock_bind_ppgtt(struct i915_address_space *vm, in mock_bind_ppgtt() argument 51 static void mock_unbind_ppgtt(struct i915_address_space *vm, in mock_unbind_ppgtt() argument 56 static void mock_cleanup(struct i915_address_space *vm) in mock_cleanup() argument 60 static void mock_clear_range(struct i915_address_space *vm, in mock_clear_range() argument 73 ppgtt->vm.gt = to_gt(i915); in mock_ppgtt() 74 ppgtt->vm.i915 = i915; in mock_ppgtt() 75 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); in mock_ppgtt() 76 ppgtt->vm in mock_ppgtt() 94 mock_bind_ggtt(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) mock_bind_ggtt() argument 102 mock_unbind_ggtt(struct i915_address_space * vm,struct i915_vma_resource * vma_res) mock_unbind_ggtt() argument [all...] |
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | gen8_ppgtt.c | 90 struct drm_i915_private *i915 = ppgtt->vm.i915; in gen8_ppgtt_notify_vgt() 91 struct intel_uncore *uncore = ppgtt->vm.gt->uncore; in gen8_ppgtt_notify_vgt() 102 if (i915_vm_is_4lvl(&ppgtt->vm)) { in gen8_ppgtt_notify_vgt() 180 static unsigned int gen8_pd_top_count(const struct i915_address_space *vm) in gen8_pd_top_count() argument 182 unsigned int shift = __gen8_pte_shift(vm->top); in gen8_pd_top_count() 184 return (vm->total + (1ull << shift) - 1) >> shift; in gen8_pd_top_count() 188 gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) in gen8_pdp_for_page_index() argument 190 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); in gen8_pdp_for_page_index() 192 if (vm->top == 2) in gen8_pdp_for_page_index() 195 return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm in gen8_pdp_for_page_index() 199 gen8_pdp_for_page_address(struct i915_address_space * const vm,const u64 addr) gen8_pdp_for_page_address() argument 204 __gen8_ppgtt_cleanup(struct i915_address_space * vm,struct i915_page_directory * pd,int count,int lvl) __gen8_ppgtt_cleanup() argument 222 gen8_ppgtt_cleanup(struct i915_address_space * vm) gen8_ppgtt_cleanup() argument 239 __gen8_ppgtt_clear(struct i915_address_space * const vm,struct i915_page_directory * const pd,u64 start,const u64 end,int lvl) __gen8_ppgtt_clear() argument 307 gen8_ppgtt_clear(struct i915_address_space * vm,u64 start,u64 length) gen8_ppgtt_clear() argument 322 __gen8_ppgtt_alloc(struct i915_address_space * const vm,struct i915_vm_pt_stash * stash,struct i915_page_directory * const pd,u64 * const start,const u64 end,int lvl) __gen8_ppgtt_alloc() argument 390 gen8_ppgtt_alloc(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,u64 start,u64 length) gen8_ppgtt_alloc() argument 406 __gen8_ppgtt_foreach(struct i915_address_space * vm,struct i915_page_directory * pd,u64 * start,u64 end,int lvl,void (* fn)(struct i915_address_space * vm,struct i915_page_table * pt,void * data),void * data) __gen8_ppgtt_foreach() argument 409 __gen8_ppgtt_foreach(struct i915_address_space * vm,struct i915_page_directory * pd,u64 * start,u64 end,int lvl,void (* fn)(struct i915_address_space * vm,struct i915_page_table * pt,void * data),void * data) __gen8_ppgtt_foreach() argument 439 gen8_ppgtt_foreach(struct i915_address_space * vm,u64 start,u64 length,void (* fn)(struct i915_address_space * vm,struct i915_page_table * pt,void * data),void * data) gen8_ppgtt_foreach() argument 441 gen8_ppgtt_foreach(struct i915_address_space * vm,u64 start,u64 length,void (* fn)(struct i915_address_space * vm,struct i915_page_table * pt,void * data),void * data) gen8_ppgtt_foreach() argument 503 xehp_ppgtt_insert_huge(struct i915_address_space * vm,struct i915_vma_resource * vma_res,struct sgt_dma * iter,unsigned int pat_index,u32 flags) xehp_ppgtt_insert_huge() argument 611 gen8_ppgtt_insert_huge(struct i915_address_space * vm,struct i915_vma_resource * vma_res,struct sgt_dma * iter,unsigned int pat_index,u32 flags) gen8_ppgtt_insert_huge() argument 735 gen8_ppgtt_insert(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) gen8_ppgtt_insert() argument 763 gen8_ppgtt_insert_entry(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) gen8_ppgtt_insert_entry() argument 784 xehp_ppgtt_insert_entry_lm(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) xehp_ppgtt_insert_entry_lm() argument 813 xehp_ppgtt_insert_entry(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) xehp_ppgtt_insert_entry() argument 826 gen8_init_scratch(struct i915_address_space * vm) gen8_init_scratch() argument 894 struct i915_address_space *vm = &ppgtt->vm; gen8_preallocate_top_level_pdp() local 925 gen8_alloc_top_pd(struct i915_address_space * vm) gen8_alloc_top_pd() argument 957 gen8_init_rsvd(struct i915_address_space * vm) gen8_init_rsvd() argument [all...] |
| H A D | gen6_ppgtt.c | 23 dma_addr_t addr = pt ? px_dma(pt) : px_dma(ppgtt->base.vm.scratch[1]); in gen6_write_pde() 74 static void gen6_ppgtt_clear_range(struct i915_address_space *vm, in gen6_ppgtt_clear_range() argument 77 struct gen6_ppgtt * const ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); in gen6_ppgtt_clear_range() 79 const gen6_pte_t scratch_pte = vm->scratch[0]->encode; in gen6_ppgtt_clear_range() 110 static void gen6_ppgtt_insert_entries(struct i915_address_space *vm, in gen6_ppgtt_insert_entries() argument 115 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); in gen6_ppgtt_insert_entries() 120 const u32 pte_encode = vm->pte_encode(0, pat_index, flags); in gen6_ppgtt_insert_entries() 166 gen6_ggtt_invalidate(ppgtt->base.vm.gt->ggtt); in gen6_flush_pd() 172 static void gen6_alloc_va_range(struct i915_address_space *vm, in gen6_alloc_va_range() argument 176 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); in gen6_alloc_va_range() 221 struct i915_address_space * const vm = &ppgtt->base.vm; gen6_ppgtt_init_scratch() local 267 gen6_ppgtt_cleanup(struct i915_address_space * vm) gen6_ppgtt_cleanup() argument 280 pd_vma_bind(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 unused) pd_vma_bind() argument 296 pd_vma_unbind(struct i915_address_space * vm,struct i915_vma_resource * vma_res) pd_vma_unbind() argument [all...] |
| /linux/sound/pci/ctxfi/ |
| H A D | ctvmem.c | 26 * Find or create vm block based on requested @size. 30 get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc) in get_vm_block() argument 36 if (size > vm->size) { in get_vm_block() 42 guard(mutex)(&vm->lock); in get_vm_block() 43 list_for_each(pos, &vm->unused) { in get_vm_block() 48 if (pos == &vm->unused) in get_vm_block() 52 /* Move the vm node from unused list to used list directly */ in get_vm_block() 53 list_move(&entry->list, &vm->used); in get_vm_block() 54 vm->size -= size; in get_vm_block() 64 list_add(&block->list, &vm in get_vm_block() 72 put_vm_block(struct ct_vm * vm,struct ct_vm_block * block) put_vm_block() argument 120 ct_vm_map(struct ct_vm * vm,struct snd_pcm_substream * substream,int size) ct_vm_map() argument 148 ct_vm_unmap(struct ct_vm * vm,struct ct_vm_block * block) ct_vm_unmap() argument 160 ct_get_ptp_phys(struct ct_vm * vm,int index) ct_get_ptp_phys() argument 167 struct ct_vm *vm; ct_vm_create() local 211 ct_vm_destroy(struct ct_vm * vm) ct_vm_destroy() argument [all...] |
| /linux/drivers/gpu/drm/i915/display/ |
| H A D | intel_dpt.c | 21 struct i915_address_space vm; member 28 #define i915_is_dpt(vm) ((vm)->is_dpt) argument 31 i915_vm_to_dpt(struct i915_address_space *vm) in i915_vm_to_dpt() argument 33 BUILD_BUG_ON(offsetof(struct i915_dpt, vm)); in i915_vm_to_dpt() 34 drm_WARN_ON(&vm->i915->drm, !i915_is_dpt(vm)); in i915_vm_to_dpt() 35 return container_of(vm, struct i915_dpt, vm); in i915_vm_to_dpt() 43 static void dpt_insert_page(struct i915_address_space *vm, in dpt_insert_page() argument 56 dpt_insert_entries(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) dpt_insert_entries() argument 78 dpt_clear_range(struct i915_address_space * vm,u64 start,u64 length) dpt_clear_range() argument 83 dpt_bind_vma(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) dpt_bind_vma() argument 113 dpt_unbind_vma(struct i915_address_space * vm,struct i915_vma_resource * vma_res) dpt_unbind_vma() argument 119 dpt_cleanup(struct i915_address_space * vm) dpt_cleanup() argument 126 intel_dpt_pin_to_ggtt(struct i915_address_space * vm,unsigned int alignment) intel_dpt_pin_to_ggtt() argument 179 intel_dpt_unpin_from_ggtt(struct i915_address_space * vm) intel_dpt_unpin_from_ggtt() argument 251 struct i915_address_space *vm; intel_dpt_create() local 315 intel_dpt_destroy(struct i915_address_space * vm) intel_dpt_destroy() argument [all...] |
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | kvm_util.h | 63 struct kvm_vm *vm; member 160 #define kvm_for_each_vcpu(vm, i, vcpu) \ in vm_get_mem_region() 161 for ((i) = 0; (i) <= (vm)->last_vcpu_id; (i)++) \ in vm_get_mem_region() 162 if (!((vcpu) = vm->vcpus[i])) \ in vm_get_mem_region() 167 memslot2region(struct kvm_vm *vm, uint32_t memslot); 169 static inline struct userspace_mem_region *vm_get_mem_region(struct kvm_vm *vm, 173 return memslot2region(vm, vm->memslots[type]); 336 static __always_inline void static_assert_is_vm(struct kvm_vm *vm) { } 338 #define __vm_ioctl(vm, cm 148 kvm_for_each_vcpu(vm,i,vcpu) global() argument 157 vm_get_mem_region(struct kvm_vm * vm,enum kvm_mem_region_type type) vm_get_mem_region() argument 313 static_assert_is_vm(struct kvm_vm * vm) static_assert_is_vm() argument 315 __vm_ioctl(vm,cmd,arg) global() argument 329 __TEST_ASSERT_VM_VCPU_IOCTL(cond,name,ret,vm) global() argument 347 TEST_ASSERT_VM_VCPU_IOCTL(cond,cmd,ret,vm) global() argument 350 vm_ioctl(vm,cmd,arg) global() argument 376 vm_check_cap(struct kvm_vm * vm,long cap) vm_check_cap() argument 384 __vm_enable_cap(struct kvm_vm * vm,uint32_t cap,uint64_t arg0) __vm_enable_cap() argument 390 vm_enable_cap(struct kvm_vm * vm,uint32_t cap,uint64_t arg0) vm_enable_cap() argument 397 vm_set_memory_attributes(struct kvm_vm * vm,uint64_t gpa,uint64_t size,uint64_t attributes) vm_set_memory_attributes() argument 418 vm_mem_set_private(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_mem_set_private() argument 424 vm_mem_set_shared(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_mem_set_shared() argument 433 vm_guest_mem_punch_hole(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_guest_mem_punch_hole() argument 439 vm_guest_mem_allocate(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_guest_mem_allocate() argument 456 kvm_vm_get_dirty_log(struct kvm_vm * vm,int slot,void * log) kvm_vm_get_dirty_log() argument 463 kvm_vm_clear_dirty_log(struct kvm_vm * vm,int slot,void * log,uint64_t first_page,uint32_t num_pages) kvm_vm_clear_dirty_log() argument 476 kvm_vm_reset_dirty_ring(struct kvm_vm * vm) kvm_vm_reset_dirty_ring() argument 481 kvm_vm_register_coalesced_io(struct kvm_vm * vm,uint64_t address,uint64_t size,bool pio) kvm_vm_register_coalesced_io() argument 494 kvm_vm_unregister_coalesced_io(struct kvm_vm * vm,uint64_t address,uint64_t size,bool pio) kvm_vm_unregister_coalesced_io() argument 507 vm_get_stats_fd(struct kvm_vm * vm) vm_get_stats_fd() argument 515 __kvm_irqfd(struct kvm_vm * vm,uint32_t gsi,int eventfd,uint32_t flags) __kvm_irqfd() argument 528 kvm_irqfd(struct kvm_vm * vm,uint32_t gsi,int eventfd,uint32_t flags) kvm_irqfd() argument 536 kvm_assign_irqfd(struct kvm_vm * vm,uint32_t gsi,int eventfd) kvm_assign_irqfd() argument 541 kvm_deassign_irqfd(struct kvm_vm * vm,uint32_t gsi,int eventfd) kvm_deassign_irqfd() argument 604 vm_get_stat(vm,stat) global() argument 644 __vm_create_guest_memfd(struct kvm_vm * vm,uint64_t size,uint64_t flags) __vm_create_guest_memfd() argument 655 vm_create_guest_memfd(struct kvm_vm * vm,uint64_t size,uint64_t flags) vm_create_guest_memfd() argument 684 vm_arch_has_protected_memory(struct kvm_vm * vm) vm_arch_has_protected_memory() argument 719 vm_untag_gpa(struct kvm_vm * vm,vm_paddr_t gpa) vm_untag_gpa() argument 928 kvm_create_device(struct kvm_vm * vm,uint64_t type) kvm_create_device() argument 977 vm_phy_pages_alloc(struct kvm_vm * vm,size_t num,vm_paddr_t paddr_min,uint32_t memslot) vm_phy_pages_alloc() argument 1114 sync_global_to_guest(vm,g) global() argument 1119 sync_global_from_guest(vm,g) global() argument 1130 write_guest_global(vm,g,val) global() argument 1158 vm_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id,void * guest_code) vm_vcpu_add() argument 1171 vm_vcpu_recreate(struct kvm_vm * vm,uint32_t vcpu_id) vm_vcpu_recreate() argument 1181 virt_pgd_alloc(struct kvm_vm * vm) virt_pgd_alloc() argument 1204 virt_pg_map(struct kvm_vm * vm,uint64_t vaddr,uint64_t paddr) virt_pg_map() argument 1228 addr_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_gva2gpa() argument 1250 virt_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent) virt_dump() argument 1256 __vm_disable_nx_huge_pages(struct kvm_vm * vm) __vm_disable_nx_huge_pages() argument [all...] |
| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_vm.c | 78 * radeon_vm_manager_init - init the vm manager 82 * Init the vm manager (cayman+). 100 * radeon_vm_manager_fini - tear down the vm manager 120 * radeon_vm_get_bos - add the vm BOs to a validation list 123 * @vm: vm providing the BOs 130 struct radeon_vm *vm, in radeon_vm_get_bos() argument 136 list = kvmalloc_objs(struct radeon_bo_list, vm->max_pde_used + 2); in radeon_vm_get_bos() 140 /* add the vm page table to the list */ in radeon_vm_get_bos() 141 list[0].robj = vm in radeon_vm_get_bos() 177 radeon_vm_grab_id(struct radeon_device * rdev,struct radeon_vm * vm,int ring) radeon_vm_grab_id() argument 236 radeon_vm_flush(struct radeon_device * rdev,struct radeon_vm * vm,int ring,struct radeon_fence * updates) radeon_vm_flush() argument 268 radeon_vm_fence(struct radeon_device * rdev,struct radeon_vm * vm,struct radeon_fence * fence) radeon_vm_fence() argument 292 radeon_vm_bo_find(struct radeon_vm * vm,struct radeon_bo * bo) radeon_vm_bo_find() argument 319 radeon_vm_bo_add(struct radeon_device * rdev,struct radeon_vm * vm,struct radeon_bo * bo) radeon_vm_bo_add() argument 451 struct radeon_vm *vm = bo_va->vm; radeon_vm_bo_set_addr() local 639 radeon_vm_update_page_directory(struct radeon_device * rdev,struct radeon_vm * vm) radeon_vm_update_page_directory() argument 813 radeon_vm_update_ptes(struct radeon_device * rdev,struct radeon_vm * vm,struct radeon_ib * ib,uint64_t start,uint64_t end,uint64_t dst,uint32_t flags) radeon_vm_update_ptes() argument 884 radeon_vm_fence_pts(struct radeon_vm * vm,uint64_t start,uint64_t end,struct radeon_fence * fence) radeon_vm_fence_pts() argument 913 struct radeon_vm *vm = bo_va->vm; radeon_vm_bo_update() local 1046 radeon_vm_clear_freed(struct radeon_device * rdev,struct radeon_vm * vm) radeon_vm_clear_freed() argument 1084 radeon_vm_clear_invalids(struct radeon_device * rdev,struct radeon_vm * vm) radeon_vm_clear_invalids() argument 1119 struct radeon_vm *vm = bo_va->vm; radeon_vm_bo_rmv() local 1171 radeon_vm_init(struct radeon_device * rdev,struct radeon_vm * vm) radeon_vm_init() argument 1231 radeon_vm_fini(struct radeon_device * rdev,struct radeon_vm * vm) radeon_vm_fini() argument [all...] |
| /linux/tools/testing/selftests/kvm/lib/riscv/ |
| H A D | processor.c | 30 static uint64_t pte_addr(struct kvm_vm *vm, uint64_t entry) in page_align() 36 static uint64_t ptrs_per_pte(struct kvm_vm *vm) in pte_addr() 55 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva, int level) 59 TEST_ASSERT(level < vm->mmu.pgtable_levels, in pte_index() argument 65 void virt_arch_pgd_alloc(struct kvm_vm *vm) in pte_index() 67 size_t nr_pages = vm_page_align(vm, ptrs_per_pte(vm) * 8) / vm->page_size; in pte_index() 69 if (vm->mmu.pgd_created) in virt_arch_pgd_alloc() argument 72 vm in virt_arch_pgd_alloc() 29 page_align(struct kvm_vm * vm,uint64_t v) page_align() argument 34 pte_addr(struct kvm_vm * vm,uint64_t entry) pte_addr() argument 40 ptrs_per_pte(struct kvm_vm * vm) ptrs_per_pte() argument 82 virt_arch_pg_map(struct kvm_vm * vm,uint64_t vaddr,uint64_t paddr) virt_arch_pg_map() argument 126 addr_arch_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_arch_gva2gpa() argument 155 pte_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent,uint64_t page,int level) pte_dump() argument 177 virt_arch_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent) virt_arch_dump() argument 198 struct kvm_vm *vm = vcpu->vm; riscv_vcpu_mmu_setup() local 298 vm_arch_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id) vm_arch_vcpu_add() argument 434 vm_init_vector_tables(struct kvm_vm * vm) vm_init_vector_tables() argument 442 vm_install_exception_handler(struct kvm_vm * vm,int vector,exception_handler_fn handler) vm_install_exception_handler() argument 450 vm_install_interrupt_handler(struct kvm_vm * vm,exception_handler_fn handler) vm_install_interrupt_handler() argument [all...] |
| /linux/tools/testing/selftests/kvm/ |
| H A D | set_memory_region_test.c | 113 struct kvm_vm *vm; in spawn_vm() local 117 vm = vm_create_with_one_vcpu(vcpu, guest_code); in spawn_vm() 119 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS_THP, in spawn_vm() 127 gpa = vm_phy_pages_alloc(vm, 2, MEM_REGION_GPA, MEM_REGION_SLOT); in spawn_vm() 130 virt_map(vm, MEM_REGION_GPA, MEM_REGION_GPA, 2); in spawn_vm() 133 hva = addr_gpa2hva(vm, MEM_REGION_GPA); in spawn_vm() 141 return vm; in spawn_vm() 182 struct kvm_vm *vm; in test_move_memory_region() local 185 vm = spawn_vm(&vcpu, &vcpu_thread, guest_code_move_memory_region); in test_move_memory_region() 188 vm_enable_cap(vm, KVM_CAP_DISABLE_QUIRKS2, KVM_X86_QUIRK_SLOT_ZAP_ALL); in test_move_memory_region() [all …]
|